[ 456.879845] env[62385]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=62385) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 456.880199] env[62385]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=62385) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 456.880243] env[62385]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=62385) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 456.880587] env[62385]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 456.976719] env[62385]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=62385) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 456.987137] env[62385]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=62385) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 457.590523] env[62385]: INFO nova.virt.driver [None req-f58ba585-75df-443a-a7ca-6eeafe7f53f1 None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 457.661861] env[62385]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 457.662036] env[62385]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 457.662166] env[62385]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=62385) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 460.699989] env[62385]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-62215ee1-6bb6-4f99-a674-47d88ab62c9c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 460.715627] env[62385]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=62385) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 460.715819] env[62385]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-549829fb-aa4e-41bb-b08f-ceb94b0b83d1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 460.748211] env[62385]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 9beb3. [ 460.748367] env[62385]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.086s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 460.748829] env[62385]: INFO nova.virt.vmwareapi.driver [None req-f58ba585-75df-443a-a7ca-6eeafe7f53f1 None None] VMware vCenter version: 7.0.3 [ 460.752158] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cdecdd9-d095-471b-a1fb-8b7984fd9378 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 460.769330] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7bd0ed0-d414-43c8-9985-2a0822b54215 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 460.775133] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47acde97-6910-48b6-bfc3-06261241e26a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 460.781571] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dae5db97-e4c2-465a-902e-c87dca062b42 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 460.794308] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-639286dd-2b39-42b4-8a69-915a542f92cf {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 460.800252] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-496ca899-7155-4eae-95b2-45d7f4ed7aca {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 460.829455] env[62385]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-0f889fb7-06e2-4230-a668-9d2a8485677e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 460.834357] env[62385]: DEBUG nova.virt.vmwareapi.driver [None req-f58ba585-75df-443a-a7ca-6eeafe7f53f1 None None] Extension org.openstack.compute already exists. {{(pid=62385) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:225}} [ 460.837088] env[62385]: INFO nova.compute.provider_config [None req-f58ba585-75df-443a-a7ca-6eeafe7f53f1 None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 461.341066] env[62385]: DEBUG nova.context [None req-f58ba585-75df-443a-a7ca-6eeafe7f53f1 None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),13828d4f-3859-464d-bbd7-665b1e575363(cell1) {{(pid=62385) load_cells /opt/stack/nova/nova/context.py:464}} [ 461.343164] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 461.343465] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 461.344172] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 461.344651] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] Acquiring lock "13828d4f-3859-464d-bbd7-665b1e575363" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 461.344884] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] Lock "13828d4f-3859-464d-bbd7-665b1e575363" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 461.345991] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] Lock "13828d4f-3859-464d-bbd7-665b1e575363" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 461.366569] env[62385]: INFO dbcounter [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] Registered counter for database nova_cell0 [ 461.374962] env[62385]: INFO dbcounter [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] Registered counter for database nova_cell1 [ 461.378468] env[62385]: DEBUG oslo_db.sqlalchemy.engines [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62385) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 461.378805] env[62385]: DEBUG oslo_db.sqlalchemy.engines [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62385) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 461.383553] env[62385]: ERROR nova.db.main.api [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 461.383553] env[62385]: result = function(*args, **kwargs) [ 461.383553] env[62385]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 461.383553] env[62385]: return func(*args, **kwargs) [ 461.383553] env[62385]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 461.383553] env[62385]: result = fn(*args, **kwargs) [ 461.383553] env[62385]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 461.383553] env[62385]: return f(*args, **kwargs) [ 461.383553] env[62385]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 461.383553] env[62385]: return db.service_get_minimum_version(context, binaries) [ 461.383553] env[62385]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 461.383553] env[62385]: _check_db_access() [ 461.383553] env[62385]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 461.383553] env[62385]: stacktrace = ''.join(traceback.format_stack()) [ 461.383553] env[62385]: [ 461.384370] env[62385]: ERROR nova.db.main.api [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 461.384370] env[62385]: result = function(*args, **kwargs) [ 461.384370] env[62385]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 461.384370] env[62385]: return func(*args, **kwargs) [ 461.384370] env[62385]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 461.384370] env[62385]: result = fn(*args, **kwargs) [ 461.384370] env[62385]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 461.384370] env[62385]: return f(*args, **kwargs) [ 461.384370] env[62385]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 461.384370] env[62385]: return db.service_get_minimum_version(context, binaries) [ 461.384370] env[62385]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 461.384370] env[62385]: _check_db_access() [ 461.384370] env[62385]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 461.384370] env[62385]: stacktrace = ''.join(traceback.format_stack()) [ 461.384370] env[62385]: [ 461.384748] env[62385]: WARNING nova.objects.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 461.384907] env[62385]: WARNING nova.objects.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] Failed to get minimum service version for cell 13828d4f-3859-464d-bbd7-665b1e575363 [ 461.385346] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] Acquiring lock "singleton_lock" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 461.385507] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] Acquired lock "singleton_lock" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 461.385769] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] Releasing lock "singleton_lock" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 461.386109] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] Full set of CONF: {{(pid=62385) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 461.386255] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] ******************************************************************************** {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 461.386383] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] Configuration options gathered from: {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 461.386516] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 461.386699] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 461.386828] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] ================================================================================ {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 461.387074] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] allow_resize_to_same_host = True {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.387249] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] arq_binding_timeout = 300 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.387381] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] backdoor_port = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.387508] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] backdoor_socket = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.387673] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] block_device_allocate_retries = 60 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.387838] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] block_device_allocate_retries_interval = 3 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.388010] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cert = self.pem {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.388184] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.388354] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] compute_monitors = [] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.388521] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] config_dir = [] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.388692] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] config_drive_format = iso9660 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.388827] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.388989] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] config_source = [] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.389169] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] console_host = devstack {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.389334] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] control_exchange = nova {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.389492] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cpu_allocation_ratio = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.389653] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] daemon = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.389820] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] debug = True {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.389978] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] default_access_ip_network_name = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.390175] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] default_availability_zone = nova {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.390336] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] default_ephemeral_format = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.390498] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] default_green_pool_size = 1000 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.390734] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.390902] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] default_schedule_zone = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.391078] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] disk_allocation_ratio = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.391245] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] enable_new_services = True {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.391421] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] enabled_apis = ['osapi_compute'] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.391585] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] enabled_ssl_apis = [] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.391747] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] flat_injected = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.391904] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] force_config_drive = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.392076] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] force_raw_images = True {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.392248] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] graceful_shutdown_timeout = 5 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.392410] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] heal_instance_info_cache_interval = 60 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.392620] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] host = cpu-1 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.392790] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.392953] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] initial_disk_allocation_ratio = 1.0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.393126] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] initial_ram_allocation_ratio = 1.0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.393338] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.393503] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] instance_build_timeout = 0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.393664] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] instance_delete_interval = 300 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.393835] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] instance_format = [instance: %(uuid)s] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.394006] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] instance_name_template = instance-%08x {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.394197] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] instance_usage_audit = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.394368] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] instance_usage_audit_period = month {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.394550] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.394726] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] instances_path = /opt/stack/data/nova/instances {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.394897] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] internal_service_availability_zone = internal {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.395075] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] key = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.395242] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] live_migration_retry_count = 30 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.395410] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] log_color = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.395591] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] log_config_append = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.395771] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.395931] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] log_dir = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.396101] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] log_file = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.396231] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] log_options = True {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.396395] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] log_rotate_interval = 1 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.396565] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] log_rotate_interval_type = days {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.396730] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] log_rotation_type = none {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.396871] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.397025] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.397201] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.397364] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.397492] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.397653] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] long_rpc_timeout = 1800 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.397815] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] max_concurrent_builds = 10 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.397972] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] max_concurrent_live_migrations = 1 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.398142] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] max_concurrent_snapshots = 5 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.398297] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] max_local_block_devices = 3 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.398451] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] max_logfile_count = 30 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.398608] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] max_logfile_size_mb = 200 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.398763] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] maximum_instance_delete_attempts = 5 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.398928] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] metadata_listen = 0.0.0.0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.399116] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] metadata_listen_port = 8775 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.399289] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] metadata_workers = 2 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.399450] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] migrate_max_retries = -1 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.399613] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] mkisofs_cmd = genisoimage {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.399816] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] my_block_storage_ip = 10.180.1.21 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.399946] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] my_ip = 10.180.1.21 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.400119] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] network_allocate_retries = 0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.400295] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.400458] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] osapi_compute_listen = 0.0.0.0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.400618] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] osapi_compute_listen_port = 8774 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.400783] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] osapi_compute_unique_server_name_scope = {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.400947] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] osapi_compute_workers = 2 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.401118] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] password_length = 12 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.401276] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] periodic_enable = True {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.401433] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] periodic_fuzzy_delay = 60 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.401597] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] pointer_model = usbtablet {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.401763] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] preallocate_images = none {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.401919] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] publish_errors = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.402061] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] pybasedir = /opt/stack/nova {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.402220] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] ram_allocation_ratio = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.402375] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] rate_limit_burst = 0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.402541] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] rate_limit_except_level = CRITICAL {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.402698] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] rate_limit_interval = 0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.402853] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] reboot_timeout = 0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.403014] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] reclaim_instance_interval = 0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.403193] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] record = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.403357] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] reimage_timeout_per_gb = 60 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.403519] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] report_interval = 120 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.403675] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] rescue_timeout = 0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.403833] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] reserved_host_cpus = 0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.403987] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] reserved_host_disk_mb = 0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.404156] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] reserved_host_memory_mb = 512 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.404314] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] reserved_huge_pages = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.404471] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] resize_confirm_window = 0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.404658] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] resize_fs_using_block_device = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.404826] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] resume_guests_state_on_host_boot = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.404994] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.405169] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] rpc_response_timeout = 60 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.405327] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] run_external_periodic_tasks = True {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.405490] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] running_deleted_instance_action = reap {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.405674] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] running_deleted_instance_poll_interval = 1800 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.405837] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] running_deleted_instance_timeout = 0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.405995] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] scheduler_instance_sync_interval = 120 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.406176] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] service_down_time = 720 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.406344] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] servicegroup_driver = db {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.406499] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] shell_completion = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.406656] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] shelved_offload_time = 0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.406816] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] shelved_poll_interval = 3600 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.407009] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] shutdown_timeout = 0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.407181] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] source_is_ipv6 = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.407339] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] ssl_only = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.407575] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.407741] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] sync_power_state_interval = 600 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.407900] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] sync_power_state_pool_size = 1000 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.408090] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] syslog_log_facility = LOG_USER {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.408254] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] tempdir = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.408415] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] timeout_nbd = 10 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.408583] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] transport_url = **** {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.408743] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] update_resources_interval = 0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.408899] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] use_cow_images = True {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.409067] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] use_eventlog = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.409227] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] use_journal = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.409384] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] use_json = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.409540] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] use_rootwrap_daemon = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.409697] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] use_stderr = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.409851] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] use_syslog = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.410008] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vcpu_pin_set = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.410180] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vif_plugging_is_fatal = True {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.410341] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vif_plugging_timeout = 300 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.410502] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] virt_mkfs = [] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.410660] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] volume_usage_poll_interval = 0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.410820] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] watch_log_file = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.410983] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] web = /usr/share/spice-html5 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 461.411178] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_concurrency.disable_process_locking = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.411715] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.411911] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.412106] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.412289] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.412473] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.412641] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.412826] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api.auth_strategy = keystone {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.412995] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api.compute_link_prefix = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.413192] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.413367] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api.dhcp_domain = novalocal {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.413537] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api.enable_instance_password = True {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.413704] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api.glance_link_prefix = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.413869] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.414052] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.414225] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api.instance_list_per_project_cells = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.414391] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api.list_records_by_skipping_down_cells = True {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.414573] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api.local_metadata_per_cell = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.414762] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api.max_limit = 1000 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.414936] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api.metadata_cache_expiration = 15 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.415125] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api.neutron_default_tenant_id = default {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.415301] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api.response_validation = warn {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.415470] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api.use_neutron_default_nets = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.415662] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.415836] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.416014] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.416210] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.416383] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api.vendordata_dynamic_targets = [] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.416549] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api.vendordata_jsonfile_path = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.416730] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.416945] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.backend = dogpile.cache.memcached {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.417140] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.backend_argument = **** {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.417315] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.config_prefix = cache.oslo {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.417485] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.dead_timeout = 60.0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.417653] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.debug_cache_backend = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.417816] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.enable_retry_client = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.417978] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.enable_socket_keepalive = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.418161] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.enabled = True {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.418326] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.enforce_fips_mode = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.418490] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.expiration_time = 600 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.418652] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.hashclient_retry_attempts = 2 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.418817] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.hashclient_retry_delay = 1.0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.418980] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.memcache_dead_retry = 300 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.419155] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.memcache_password = **** {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.419319] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.419482] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.419646] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.memcache_pool_maxsize = 10 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.419810] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.419973] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.memcache_sasl_enabled = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.420184] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.420358] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.memcache_socket_timeout = 1.0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.420523] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.memcache_username = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.420685] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.proxies = [] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.420850] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.redis_db = 0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.421023] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.redis_password = **** {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.421191] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.redis_sentinel_service_name = mymaster {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.421365] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.421534] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.redis_server = localhost:6379 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.421698] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.redis_socket_timeout = 1.0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.421859] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.redis_username = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.422030] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.retry_attempts = 2 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.422201] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.retry_delay = 0.0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.422364] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.socket_keepalive_count = 1 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.422531] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.socket_keepalive_idle = 1 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.422723] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.socket_keepalive_interval = 1 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.422884] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.tls_allowed_ciphers = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.423053] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.tls_cafile = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.423213] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.tls_certfile = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.423376] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.tls_enabled = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.423533] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cache.tls_keyfile = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.423701] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cinder.auth_section = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.423874] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cinder.auth_type = password {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.424054] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cinder.cafile = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.424240] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cinder.catalog_info = volumev3::publicURL {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.424400] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cinder.certfile = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.424588] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cinder.collect_timing = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.424769] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cinder.cross_az_attach = True {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.424935] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cinder.debug = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.425110] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cinder.endpoint_template = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.425277] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cinder.http_retries = 3 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.425438] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cinder.insecure = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.425618] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cinder.keyfile = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.425800] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cinder.os_region_name = RegionOne {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.425964] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cinder.split_loggers = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.426139] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cinder.timeout = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.426314] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.426475] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] compute.cpu_dedicated_set = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.426634] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] compute.cpu_shared_set = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.426800] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] compute.image_type_exclude_list = [] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.426991] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.427176] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] compute.max_concurrent_disk_ops = 0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.427339] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] compute.max_disk_devices_to_attach = -1 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.427500] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.427669] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.427838] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] compute.resource_provider_association_refresh = 300 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.427999] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.428175] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] compute.shutdown_retry_interval = 10 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.428353] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.428531] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] conductor.workers = 2 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.428709] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] console.allowed_origins = [] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.428871] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] console.ssl_ciphers = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.429063] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] console.ssl_minimum_version = default {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.429242] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] consoleauth.enforce_session_timeout = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.429414] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] consoleauth.token_ttl = 600 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.429583] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cyborg.cafile = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.429748] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cyborg.certfile = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.429906] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cyborg.collect_timing = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.430077] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cyborg.connect_retries = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.430238] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cyborg.connect_retry_delay = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.430396] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cyborg.endpoint_override = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.430559] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cyborg.insecure = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.430719] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cyborg.keyfile = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.430878] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cyborg.max_version = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.431052] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cyborg.min_version = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.431215] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cyborg.region_name = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.431370] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cyborg.retriable_status_codes = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.431528] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cyborg.service_name = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.431696] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cyborg.service_type = accelerator {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.431855] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cyborg.split_loggers = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.432019] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cyborg.status_code_retries = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.432177] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cyborg.status_code_retry_delay = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.432335] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cyborg.timeout = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.432512] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.432673] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] cyborg.version = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.432855] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] database.backend = sqlalchemy {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.433039] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] database.connection = **** {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.433216] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] database.connection_debug = 0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.433387] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] database.connection_parameters = {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.433552] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] database.connection_recycle_time = 3600 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.433717] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] database.connection_trace = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.433879] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] database.db_inc_retry_interval = True {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.434051] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] database.db_max_retries = 20 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.434217] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] database.db_max_retry_interval = 10 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.434378] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] database.db_retry_interval = 1 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.434555] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] database.max_overflow = 50 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.434734] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] database.max_pool_size = 5 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.434899] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] database.max_retries = 10 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.435080] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.435243] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] database.mysql_wsrep_sync_wait = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.435401] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] database.pool_timeout = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.435580] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] database.retry_interval = 10 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.435754] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] database.slave_connection = **** {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.435916] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] database.sqlite_synchronous = True {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.436089] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] database.use_db_reconnect = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.436271] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api_database.backend = sqlalchemy {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.436440] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api_database.connection = **** {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.436605] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api_database.connection_debug = 0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.436775] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api_database.connection_parameters = {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.436963] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api_database.connection_recycle_time = 3600 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.437150] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api_database.connection_trace = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.437316] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api_database.db_inc_retry_interval = True {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.437481] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api_database.db_max_retries = 20 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.437644] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api_database.db_max_retry_interval = 10 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.437808] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api_database.db_retry_interval = 1 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.437968] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api_database.max_overflow = 50 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.438157] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api_database.max_pool_size = 5 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.438320] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api_database.max_retries = 10 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.438496] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.438667] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.438830] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api_database.pool_timeout = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.438993] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api_database.retry_interval = 10 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.439168] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api_database.slave_connection = **** {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.439329] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] api_database.sqlite_synchronous = True {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.439502] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] devices.enabled_mdev_types = [] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.439679] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.439854] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] ephemeral_storage_encryption.default_format = luks {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.440027] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] ephemeral_storage_encryption.enabled = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.440194] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.440365] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] glance.api_servers = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.440529] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] glance.cafile = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.440692] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] glance.certfile = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.440855] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] glance.collect_timing = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.441056] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] glance.connect_retries = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.441318] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] glance.connect_retry_delay = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.441498] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] glance.debug = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.441686] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] glance.default_trusted_certificate_ids = [] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.441854] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] glance.enable_certificate_validation = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.443539] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] glance.enable_rbd_download = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.443818] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] glance.endpoint_override = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.444041] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] glance.insecure = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.451733] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] glance.keyfile = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.451733] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] glance.max_version = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.451733] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] glance.min_version = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.451733] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] glance.num_retries = 3 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.451733] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] glance.rbd_ceph_conf = {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.451733] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] glance.rbd_connect_timeout = 5 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.451733] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] glance.rbd_pool = {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.451998] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] glance.rbd_user = {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.451998] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] glance.region_name = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.451998] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] glance.retriable_status_codes = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.451998] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] glance.service_name = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.451998] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] glance.service_type = image {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.451998] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] glance.split_loggers = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.451998] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] glance.status_code_retries = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.452208] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] glance.status_code_retry_delay = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.452208] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] glance.timeout = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.452208] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.452208] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] glance.verify_glance_signatures = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.452208] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] glance.version = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.452208] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] guestfs.debug = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.452208] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] mks.enabled = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.452388] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.452388] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] image_cache.manager_interval = 2400 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.452388] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] image_cache.precache_concurrency = 1 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.452388] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] image_cache.remove_unused_base_images = True {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.452388] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.452388] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.452388] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] image_cache.subdirectory_name = _base {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.452566] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] ironic.api_max_retries = 60 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.452566] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] ironic.api_retry_interval = 2 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.452566] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] ironic.auth_section = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.452566] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] ironic.auth_type = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.452566] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] ironic.cafile = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.452566] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] ironic.certfile = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.452566] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] ironic.collect_timing = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.452741] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] ironic.conductor_group = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.452741] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] ironic.connect_retries = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.452741] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] ironic.connect_retry_delay = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.452741] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] ironic.endpoint_override = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.452741] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] ironic.insecure = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.452741] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] ironic.keyfile = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.452741] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] ironic.max_version = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.452909] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] ironic.min_version = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.452909] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] ironic.peer_list = [] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.452909] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] ironic.region_name = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.452909] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] ironic.retriable_status_codes = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.452909] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] ironic.serial_console_state_timeout = 10 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.452909] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] ironic.service_name = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.452909] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] ironic.service_type = baremetal {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.453129] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] ironic.shard = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.453129] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] ironic.split_loggers = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.453129] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] ironic.status_code_retries = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.453129] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] ironic.status_code_retry_delay = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.453280] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] ironic.timeout = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.453420] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.454128] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] ironic.version = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.454128] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.454128] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] key_manager.fixed_key = **** {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.454128] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.454311] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] barbican.barbican_api_version = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.454427] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] barbican.barbican_endpoint = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.455215] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] barbican.barbican_endpoint_type = public {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.455215] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] barbican.barbican_region_name = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.455215] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] barbican.cafile = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.455215] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] barbican.certfile = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.455426] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] barbican.collect_timing = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.455539] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] barbican.insecure = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.456377] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] barbican.keyfile = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.456377] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] barbican.number_of_retries = 60 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.456377] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] barbican.retry_delay = 1 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.456377] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] barbican.send_service_user_token = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.456552] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] barbican.split_loggers = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.456624] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] barbican.timeout = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.460945] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] barbican.verify_ssl = True {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.460945] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] barbican.verify_ssl_path = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.460945] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] barbican_service_user.auth_section = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.460945] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] barbican_service_user.auth_type = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.460945] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] barbican_service_user.cafile = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.460945] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] barbican_service_user.certfile = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.460945] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] barbican_service_user.collect_timing = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.461239] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] barbican_service_user.insecure = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.461239] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] barbican_service_user.keyfile = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.461239] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] barbican_service_user.split_loggers = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.461239] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] barbican_service_user.timeout = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.461239] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vault.approle_role_id = **** {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.461239] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vault.approle_secret_id = **** {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.461239] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vault.kv_mountpoint = secret {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.461477] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vault.kv_path = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.461477] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vault.kv_version = 2 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.461477] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vault.namespace = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.461477] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vault.root_token_id = **** {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.461477] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vault.ssl_ca_crt_file = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.461477] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vault.timeout = 60.0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.461477] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vault.use_ssl = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.461700] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.461700] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] keystone.auth_section = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.461700] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] keystone.auth_type = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.461700] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] keystone.cafile = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.461700] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] keystone.certfile = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.461853] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] keystone.collect_timing = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.461894] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] keystone.connect_retries = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.462036] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] keystone.connect_retry_delay = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.463518] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] keystone.endpoint_override = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.463518] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] keystone.insecure = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.463518] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] keystone.keyfile = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.463518] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] keystone.max_version = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.463518] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] keystone.min_version = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.463518] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] keystone.region_name = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.463518] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] keystone.retriable_status_codes = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.463795] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] keystone.service_name = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.463795] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] keystone.service_type = identity {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.463795] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] keystone.split_loggers = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.463923] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] keystone.status_code_retries = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.465797] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] keystone.status_code_retry_delay = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.465797] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] keystone.timeout = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.465797] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.465797] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] keystone.version = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.465797] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.connection_uri = {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.465797] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.cpu_mode = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.465797] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.cpu_model_extra_flags = [] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.466137] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.cpu_models = [] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.466137] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.cpu_power_governor_high = performance {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.466137] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.cpu_power_governor_low = powersave {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.466137] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.cpu_power_management = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.466272] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.468579] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.device_detach_attempts = 8 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.468579] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.device_detach_timeout = 20 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.468579] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.disk_cachemodes = [] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.468579] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.disk_prefix = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.468579] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.enabled_perf_events = [] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.468579] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.file_backed_memory = 0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.468579] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.gid_maps = [] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.468896] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.hw_disk_discard = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.468896] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.hw_machine_type = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.468896] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.images_rbd_ceph_conf = {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.468896] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.468896] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.468896] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.images_rbd_glance_store_name = {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.468896] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.images_rbd_pool = rbd {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.469169] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.images_type = default {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.469268] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.images_volume_group = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.471945] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.inject_key = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.471945] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.inject_partition = -2 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.471945] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.inject_password = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.471945] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.iscsi_iface = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.471945] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.iser_use_multipath = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.471945] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.live_migration_bandwidth = 0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.471945] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.472264] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.live_migration_downtime = 500 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.472264] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.472264] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.472264] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.live_migration_inbound_addr = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.472264] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.472264] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.live_migration_permit_post_copy = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.472264] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.live_migration_scheme = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.472536] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.live_migration_timeout_action = abort {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.472536] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.live_migration_tunnelled = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.472536] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.live_migration_uri = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.472536] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.live_migration_with_native_tls = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.472536] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.max_queues = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.472772] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.472937] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.473137] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.nfs_mount_options = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.473465] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.473653] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.473850] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.num_iser_scan_tries = 5 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.474072] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.num_memory_encrypted_guests = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.474263] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.474455] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.num_pcie_ports = 0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.474683] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.num_volume_scan_tries = 5 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.474870] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.pmem_namespaces = [] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.474984] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.quobyte_client_cfg = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.475292] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.475471] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.rbd_connect_timeout = 5 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.475679] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.475867] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.476045] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.rbd_secret_uuid = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.476213] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.rbd_user = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.476378] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.476552] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.remote_filesystem_transport = ssh {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.476718] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.rescue_image_id = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.476882] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.rescue_kernel_id = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.477047] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.rescue_ramdisk_id = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.477223] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.477385] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.rx_queue_size = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.477553] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.smbfs_mount_options = {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.477831] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.478009] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.snapshot_compression = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.478179] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.snapshot_image_format = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.478398] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.478564] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.sparse_logical_volumes = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.478727] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.swtpm_enabled = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.478894] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.swtpm_group = tss {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.479083] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.swtpm_user = tss {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.479261] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.sysinfo_serial = unique {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.479422] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.tb_cache_size = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.479580] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.tx_queue_size = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.479782] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.uid_maps = [] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.479953] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.use_virtio_for_bridges = True {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.480139] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.virt_type = kvm {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.480311] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.volume_clear = zero {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.480477] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.volume_clear_size = 0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.480646] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.volume_use_multipath = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.480807] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.vzstorage_cache_path = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.480974] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.481154] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.vzstorage_mount_group = qemu {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.481320] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.vzstorage_mount_opts = [] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.481484] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.481759] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.481938] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.vzstorage_mount_user = stack {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.482118] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.482297] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] neutron.auth_section = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.482471] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] neutron.auth_type = password {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.482634] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] neutron.cafile = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.482799] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] neutron.certfile = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.482962] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] neutron.collect_timing = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.483134] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] neutron.connect_retries = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.483297] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] neutron.connect_retry_delay = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.483468] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] neutron.default_floating_pool = public {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.483628] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] neutron.endpoint_override = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.483803] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] neutron.extension_sync_interval = 600 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.483967] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] neutron.http_retries = 3 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.484155] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] neutron.insecure = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.484318] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] neutron.keyfile = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.484479] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] neutron.max_version = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.484676] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.484845] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] neutron.min_version = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.485020] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] neutron.ovs_bridge = br-int {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.485190] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] neutron.physnets = [] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.485359] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] neutron.region_name = RegionOne {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.485519] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] neutron.retriable_status_codes = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.485732] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] neutron.service_metadata_proxy = True {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.485909] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] neutron.service_name = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.486099] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] neutron.service_type = network {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.486269] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] neutron.split_loggers = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.486429] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] neutron.status_code_retries = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.486588] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] neutron.status_code_retry_delay = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.486750] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] neutron.timeout = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.486929] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.487104] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] neutron.version = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.487282] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] notifications.bdms_in_notifications = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.487457] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] notifications.default_level = INFO {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.487632] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] notifications.notification_format = unversioned {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.487798] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] notifications.notify_on_state_change = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.487975] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.488168] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] pci.alias = [] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.488339] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] pci.device_spec = [] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.488503] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] pci.report_in_placement = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.488676] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] placement.auth_section = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.488849] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] placement.auth_type = password {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.489032] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.489205] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] placement.cafile = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.489364] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] placement.certfile = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.489528] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] placement.collect_timing = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.489689] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] placement.connect_retries = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.489846] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] placement.connect_retry_delay = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.490017] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] placement.default_domain_id = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.490175] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] placement.default_domain_name = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.490333] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] placement.domain_id = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.490491] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] placement.domain_name = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.490659] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] placement.endpoint_override = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.490909] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] placement.insecure = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.491111] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] placement.keyfile = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.491279] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] placement.max_version = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.491441] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] placement.min_version = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.491612] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] placement.password = **** {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.491804] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] placement.project_domain_id = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.491983] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] placement.project_domain_name = Default {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.492175] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] placement.project_id = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.492354] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] placement.project_name = service {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.492527] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] placement.region_name = RegionOne {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.492690] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] placement.retriable_status_codes = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.492850] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] placement.service_name = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.493027] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] placement.service_type = placement {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.493196] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] placement.split_loggers = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.493356] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] placement.status_code_retries = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.493521] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] placement.status_code_retry_delay = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.493705] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] placement.system_scope = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.493885] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] placement.timeout = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.494067] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] placement.trust_id = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.494238] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] placement.user_domain_id = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.494409] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] placement.user_domain_name = Default {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.494591] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] placement.user_id = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.494775] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] placement.username = nova {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.494958] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.495134] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] placement.version = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.495315] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] quota.cores = 20 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.495481] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] quota.count_usage_from_placement = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.495653] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.495823] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] quota.injected_file_content_bytes = 10240 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.495987] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] quota.injected_file_path_length = 255 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.496166] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] quota.injected_files = 5 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.496331] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] quota.instances = 10 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.496497] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] quota.key_pairs = 100 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.496662] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] quota.metadata_items = 128 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.496830] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] quota.ram = 51200 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.496992] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] quota.recheck_quota = True {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.497172] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] quota.server_group_members = 10 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.497363] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] quota.server_groups = 10 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.497548] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.497719] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.497884] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] scheduler.image_metadata_prefilter = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.498069] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.498246] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] scheduler.max_attempts = 3 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.498413] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] scheduler.max_placement_results = 1000 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.498580] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.498745] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] scheduler.query_placement_for_image_type_support = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.498905] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.499089] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] scheduler.workers = 2 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.499272] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.499441] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.499621] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.499793] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.499958] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.500133] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.500298] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.500483] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.500652] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] filter_scheduler.host_subset_size = 1 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.500818] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.500977] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.501154] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.501319] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] filter_scheduler.isolated_hosts = [] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.501485] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] filter_scheduler.isolated_images = [] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.501647] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.501810] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.501973] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.502146] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] filter_scheduler.pci_in_placement = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.502311] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.502472] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.502634] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.502796] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.502957] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.503145] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.503310] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] filter_scheduler.track_instance_changes = True {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.503488] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.503703] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] metrics.required = True {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.503882] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] metrics.weight_multiplier = 1.0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.504076] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.504255] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] metrics.weight_setting = [] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.504576] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.504755] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] serial_console.enabled = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.504934] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] serial_console.port_range = 10000:20000 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.505118] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.505288] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.505454] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] serial_console.serialproxy_port = 6083 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.505629] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] service_user.auth_section = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.505829] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] service_user.auth_type = password {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.505994] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] service_user.cafile = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.506233] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] service_user.certfile = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.506418] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] service_user.collect_timing = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.506585] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] service_user.insecure = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.506746] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] service_user.keyfile = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.506919] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] service_user.send_service_user_token = True {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.507107] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] service_user.split_loggers = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.507279] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] service_user.timeout = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.507506] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] spice.agent_enabled = True {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.507686] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] spice.enabled = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.508040] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.508245] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.508419] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] spice.html5proxy_port = 6082 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.508582] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] spice.image_compression = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.508743] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] spice.jpeg_compression = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.508902] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] spice.playback_compression = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.509075] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] spice.require_secure = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.509248] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] spice.server_listen = 127.0.0.1 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.509419] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.509579] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] spice.streaming_mode = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.509740] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] spice.zlib_compression = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.509906] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] upgrade_levels.baseapi = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.510088] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] upgrade_levels.compute = auto {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.510252] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] upgrade_levels.conductor = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.510411] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] upgrade_levels.scheduler = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.510578] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vendordata_dynamic_auth.auth_section = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.510742] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vendordata_dynamic_auth.auth_type = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.510900] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vendordata_dynamic_auth.cafile = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.511067] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vendordata_dynamic_auth.certfile = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.511231] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.511390] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vendordata_dynamic_auth.insecure = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.511546] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vendordata_dynamic_auth.keyfile = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.511708] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.511864] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vendordata_dynamic_auth.timeout = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.512057] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vmware.api_retry_count = 10 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.512229] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vmware.ca_file = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.512401] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vmware.cache_prefix = devstack-image-cache {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.512570] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vmware.cluster_name = testcl1 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.512736] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vmware.connection_pool_size = 10 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.512893] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vmware.console_delay_seconds = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.513070] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vmware.datastore_regex = ^datastore.* {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.513277] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.513450] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vmware.host_password = **** {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.513617] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vmware.host_port = 443 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.513787] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vmware.host_username = administrator@vsphere.local {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.513956] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vmware.insecure = True {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.514131] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vmware.integration_bridge = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.514298] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vmware.maximum_objects = 100 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.514457] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vmware.pbm_default_policy = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.514649] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vmware.pbm_enabled = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.514812] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vmware.pbm_wsdl_location = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.514981] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.515153] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vmware.serial_port_proxy_uri = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.515312] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vmware.serial_port_service_uri = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.515476] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vmware.task_poll_interval = 0.5 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.515663] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vmware.use_linked_clone = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.515847] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vmware.vnc_keymap = en-us {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.516026] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vmware.vnc_port = 5900 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.516195] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vmware.vnc_port_total = 10000 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.516381] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vnc.auth_schemes = ['none'] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.516557] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vnc.enabled = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.516872] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.517067] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.517248] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vnc.novncproxy_port = 6080 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.517427] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vnc.server_listen = 127.0.0.1 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.517598] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.517762] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vnc.vencrypt_ca_certs = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.517920] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vnc.vencrypt_client_cert = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.518090] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vnc.vencrypt_client_key = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.518267] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.518430] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] workarounds.disable_deep_image_inspection = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.518593] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.518758] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.518920] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.519092] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] workarounds.disable_rootwrap = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.519257] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] workarounds.enable_numa_live_migration = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.519417] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.519577] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.519737] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.519895] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] workarounds.libvirt_disable_apic = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.520062] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.520228] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.520389] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.520558] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.520723] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.520880] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.521048] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.521213] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.521372] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.521534] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.521719] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.521889] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] wsgi.client_socket_timeout = 900 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.522082] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] wsgi.default_pool_size = 1000 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.522260] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] wsgi.keep_alive = True {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.522428] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] wsgi.max_header_line = 16384 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.522595] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] wsgi.secure_proxy_ssl_header = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.522757] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] wsgi.ssl_ca_file = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.522917] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] wsgi.ssl_cert_file = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.523089] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] wsgi.ssl_key_file = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.523257] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] wsgi.tcp_keepidle = 600 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.523433] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.523601] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] zvm.ca_file = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.523765] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] zvm.cloud_connector_url = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.524060] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.524237] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] zvm.reachable_timeout = 300 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.524419] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_policy.enforce_new_defaults = True {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.524852] env[62385]: WARNING oslo_config.cfg [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 461.525055] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_policy.enforce_scope = True {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.525244] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_policy.policy_default_rule = default {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.525428] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.525606] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_policy.policy_file = policy.yaml {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.525818] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.525988] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.526182] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.526346] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.526513] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.526684] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.526875] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.527106] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] profiler.connection_string = messaging:// {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.527288] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] profiler.enabled = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.527463] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] profiler.es_doc_type = notification {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.527632] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] profiler.es_scroll_size = 10000 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.527807] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] profiler.es_scroll_time = 2m {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.527972] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] profiler.filter_error_trace = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.528158] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] profiler.hmac_keys = **** {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.528328] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] profiler.sentinel_service_name = mymaster {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.528496] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] profiler.socket_timeout = 0.1 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.528659] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] profiler.trace_requests = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.528820] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] profiler.trace_sqlalchemy = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.529006] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] profiler_jaeger.process_tags = {} {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.529174] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] profiler_jaeger.service_name_prefix = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.529336] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] profiler_otlp.service_name_prefix = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.529502] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] remote_debug.host = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.529661] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] remote_debug.port = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.529840] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.530006] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.530193] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.530358] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.530520] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.530681] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.530841] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.531020] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.531180] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.531348] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.531507] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.531677] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.531847] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.532024] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.532200] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.532367] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.532532] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.532707] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.532870] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.533296] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.533296] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.533408] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.533515] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.533681] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.533849] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.534030] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.534207] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.534372] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.534550] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.534744] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_rabbit.ssl = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.534922] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.535107] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.535273] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.535444] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.535640] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_rabbit.ssl_version = {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.535822] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.536018] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.536192] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_notifications.retry = -1 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.536375] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.536551] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_messaging_notifications.transport_url = **** {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.536723] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_limit.auth_section = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.536916] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_limit.auth_type = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.537141] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_limit.cafile = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.537316] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_limit.certfile = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.537484] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_limit.collect_timing = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.537644] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_limit.connect_retries = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.537806] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_limit.connect_retry_delay = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.537963] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_limit.endpoint_id = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.538148] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_limit.endpoint_override = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.538314] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_limit.insecure = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.538471] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_limit.keyfile = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.538629] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_limit.max_version = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.538787] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_limit.min_version = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.538941] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_limit.region_name = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.539114] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_limit.retriable_status_codes = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.539274] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_limit.service_name = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.539429] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_limit.service_type = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.539591] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_limit.split_loggers = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.539750] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_limit.status_code_retries = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.539906] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_limit.status_code_retry_delay = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.540070] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_limit.timeout = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.540229] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_limit.valid_interfaces = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.540388] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_limit.version = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.540551] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_reports.file_event_handler = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.540717] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.540875] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] oslo_reports.log_dir = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.541056] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.541222] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.541379] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.541545] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.541709] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.541865] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.542053] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.542226] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vif_plug_ovs_privileged.group = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.542384] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.542550] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.542714] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.542871] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] vif_plug_ovs_privileged.user = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.543051] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] os_vif_linux_bridge.flat_interface = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.543234] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.543405] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.543575] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.543747] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.543912] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.544090] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.544255] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.544432] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.544638] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] os_vif_ovs.isolate_vif = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.544816] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.544984] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.545168] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.545338] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] os_vif_ovs.ovsdb_interface = native {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.545500] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] os_vif_ovs.per_port_bridge = False {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.545720] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] os_brick.lock_path = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.545908] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.546086] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.546263] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] privsep_osbrick.capabilities = [21] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.546426] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] privsep_osbrick.group = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.546586] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] privsep_osbrick.helper_command = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.546752] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.546936] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.547128] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] privsep_osbrick.user = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.547309] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.547471] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] nova_sys_admin.group = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.547630] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] nova_sys_admin.helper_command = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.547795] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.547957] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.548130] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] nova_sys_admin.user = None {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 461.548264] env[62385]: DEBUG oslo_service.service [None req-ec5a7f32-42c4-498f-ae82-ae8692a20a28 None None] ******************************************************************************** {{(pid=62385) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 461.548791] env[62385]: INFO nova.service [-] Starting compute node (version 30.0.1) [ 462.051972] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-5d9d4142-e3af-4601-9d0f-6e9895f543c1 None None] Getting list of instances from cluster (obj){ [ 462.051972] env[62385]: value = "domain-c8" [ 462.051972] env[62385]: _type = "ClusterComputeResource" [ 462.051972] env[62385]: } {{(pid=62385) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 462.053123] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c5edd2f-12f8-4113-8989-e177de384460 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 462.062659] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-5d9d4142-e3af-4601-9d0f-6e9895f543c1 None None] Got total of 0 instances {{(pid=62385) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 462.063316] env[62385]: WARNING nova.virt.vmwareapi.driver [None req-5d9d4142-e3af-4601-9d0f-6e9895f543c1 None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 462.063877] env[62385]: INFO nova.virt.node [None req-5d9d4142-e3af-4601-9d0f-6e9895f543c1 None None] Generated node identity 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a [ 462.064196] env[62385]: INFO nova.virt.node [None req-5d9d4142-e3af-4601-9d0f-6e9895f543c1 None None] Wrote node identity 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a to /opt/stack/data/n-cpu-1/compute_id [ 462.567314] env[62385]: WARNING nova.compute.manager [None req-5d9d4142-e3af-4601-9d0f-6e9895f543c1 None None] Compute nodes ['8c6086eb-4f96-48ed-af58-d7d7f36e7c5a'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 463.572708] env[62385]: INFO nova.compute.manager [None req-5d9d4142-e3af-4601-9d0f-6e9895f543c1 None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 464.578191] env[62385]: WARNING nova.compute.manager [None req-5d9d4142-e3af-4601-9d0f-6e9895f543c1 None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 464.578547] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5d9d4142-e3af-4601-9d0f-6e9895f543c1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 464.578672] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5d9d4142-e3af-4601-9d0f-6e9895f543c1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 464.578826] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5d9d4142-e3af-4601-9d0f-6e9895f543c1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 464.578980] env[62385]: DEBUG nova.compute.resource_tracker [None req-5d9d4142-e3af-4601-9d0f-6e9895f543c1 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62385) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 464.579951] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27d7ee1a-3426-4eaf-9583-28b3a14dd9a0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 464.588224] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd2b4f4f-6f15-452d-8e4b-964cc8519d3b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 464.601512] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a7cf06a-e093-4c33-842b-94029155c1f4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 464.607598] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c0ebfa3-e273-4d5e-b2fa-31f2fb4a6615 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 464.636443] env[62385]: DEBUG nova.compute.resource_tracker [None req-5d9d4142-e3af-4601-9d0f-6e9895f543c1 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181407MB free_disk=155GB free_vcpus=48 pci_devices=None {{(pid=62385) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 464.636592] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5d9d4142-e3af-4601-9d0f-6e9895f543c1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 464.636781] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5d9d4142-e3af-4601-9d0f-6e9895f543c1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 465.139049] env[62385]: WARNING nova.compute.resource_tracker [None req-5d9d4142-e3af-4601-9d0f-6e9895f543c1 None None] No compute node record for cpu-1:8c6086eb-4f96-48ed-af58-d7d7f36e7c5a: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a could not be found. [ 465.642836] env[62385]: INFO nova.compute.resource_tracker [None req-5d9d4142-e3af-4601-9d0f-6e9895f543c1 None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a [ 467.150499] env[62385]: DEBUG nova.compute.resource_tracker [None req-5d9d4142-e3af-4601-9d0f-6e9895f543c1 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 467.150875] env[62385]: DEBUG nova.compute.resource_tracker [None req-5d9d4142-e3af-4601-9d0f-6e9895f543c1 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 467.297661] env[62385]: INFO nova.scheduler.client.report [None req-5d9d4142-e3af-4601-9d0f-6e9895f543c1 None None] [req-b3237dec-e584-4626-ac60-59164c07751e] Created resource provider record via placement API for resource provider with UUID 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 467.314096] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6348bec-29ee-4d3a-b805-0a473cc8db9b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 467.322049] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b01b64a8-1970-4766-a6d4-d7e389aa97b3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 467.351998] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67c27dfb-223d-4d76-a830-5c9acbca5739 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 467.358800] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5fecb8b-64ef-4fbc-862f-a380d49f4f73 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 467.371577] env[62385]: DEBUG nova.compute.provider_tree [None req-5d9d4142-e3af-4601-9d0f-6e9895f543c1 None None] Updating inventory in ProviderTree for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 467.907445] env[62385]: DEBUG nova.scheduler.client.report [None req-5d9d4142-e3af-4601-9d0f-6e9895f543c1 None None] Updated inventory for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 467.907672] env[62385]: DEBUG nova.compute.provider_tree [None req-5d9d4142-e3af-4601-9d0f-6e9895f543c1 None None] Updating resource provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a generation from 0 to 1 during operation: update_inventory {{(pid=62385) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 467.907818] env[62385]: DEBUG nova.compute.provider_tree [None req-5d9d4142-e3af-4601-9d0f-6e9895f543c1 None None] Updating inventory in ProviderTree for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 467.955770] env[62385]: DEBUG nova.compute.provider_tree [None req-5d9d4142-e3af-4601-9d0f-6e9895f543c1 None None] Updating resource provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a generation from 1 to 2 during operation: update_traits {{(pid=62385) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 468.460803] env[62385]: DEBUG nova.compute.resource_tracker [None req-5d9d4142-e3af-4601-9d0f-6e9895f543c1 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62385) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 468.461242] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5d9d4142-e3af-4601-9d0f-6e9895f543c1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.824s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 468.461242] env[62385]: DEBUG nova.service [None req-5d9d4142-e3af-4601-9d0f-6e9895f543c1 None None] Creating RPC server for service compute {{(pid=62385) start /opt/stack/nova/nova/service.py:186}} [ 468.476206] env[62385]: DEBUG nova.service [None req-5d9d4142-e3af-4601-9d0f-6e9895f543c1 None None] Join ServiceGroup membership for this service compute {{(pid=62385) start /opt/stack/nova/nova/service.py:203}} [ 468.476398] env[62385]: DEBUG nova.servicegroup.drivers.db [None req-5d9d4142-e3af-4601-9d0f-6e9895f543c1 None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=62385) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 491.486028] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._sync_power_states {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 491.986209] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Getting list of instances from cluster (obj){ [ 491.986209] env[62385]: value = "domain-c8" [ 491.986209] env[62385]: _type = "ClusterComputeResource" [ 491.986209] env[62385]: } {{(pid=62385) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 491.993021] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02c991b1-a10f-40ea-a910-dc48bf07b720 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 491.997592] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Got total of 0 instances {{(pid=62385) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 491.997833] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 491.998172] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Getting list of instances from cluster (obj){ [ 491.998172] env[62385]: value = "domain-c8" [ 491.998172] env[62385]: _type = "ClusterComputeResource" [ 491.998172] env[62385]: } {{(pid=62385) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 492.002515] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7ce9136-f26e-4669-8cac-fb94f15b51d9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.011586] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Got total of 0 instances {{(pid=62385) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 506.063050] env[62385]: DEBUG oslo_concurrency.lockutils [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] Acquiring lock "adb14669-5f0b-4d3a-9cbf-293117a1d1cd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 506.063369] env[62385]: DEBUG oslo_concurrency.lockutils [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] Lock "adb14669-5f0b-4d3a-9cbf-293117a1d1cd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 506.071761] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] Acquiring lock "ad6baae6-3925-41d7-9b95-0a5fe8add72d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 506.072116] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] Lock "ad6baae6-3925-41d7-9b95-0a5fe8add72d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 506.567317] env[62385]: DEBUG nova.compute.manager [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 506.576656] env[62385]: DEBUG nova.compute.manager [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 506.789748] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] Acquiring lock "d4f4eb24-ea78-4ab1-9786-818ea7769342" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 506.792646] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] Lock "d4f4eb24-ea78-4ab1-9786-818ea7769342" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 507.120443] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 507.120673] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 507.123762] env[62385]: INFO nova.compute.claims [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 507.126661] env[62385]: DEBUG oslo_concurrency.lockutils [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 507.292952] env[62385]: DEBUG nova.compute.manager [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 507.513143] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Acquiring lock "077b25b0-3586-4c56-81fa-526ff9c0ed1f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 507.513356] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Lock "077b25b0-3586-4c56-81fa-526ff9c0ed1f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 507.824947] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 508.017926] env[62385]: DEBUG nova.compute.manager [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 508.233830] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2f8b802-656d-42f1-93fd-47e8d02cdd1d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 508.246535] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a639199d-0e05-478e-9938-4397c9cf14a0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 508.281297] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaf8b149-cd05-43ab-886d-b159ed0fef00 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 508.289875] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5592b596-69a5-4109-bb00-38d2133619e3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 508.306968] env[62385]: DEBUG nova.compute.provider_tree [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 508.553271] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 508.817938] env[62385]: DEBUG nova.scheduler.client.report [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 508.857057] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Acquiring lock "b28550f5-4f8e-43ac-a020-c025586c6688" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 508.857325] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Lock "b28550f5-4f8e-43ac-a020-c025586c6688" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 509.321447] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.200s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 509.322376] env[62385]: DEBUG nova.compute.manager [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 509.326880] env[62385]: DEBUG oslo_concurrency.lockutils [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.200s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 509.329715] env[62385]: INFO nova.compute.claims [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 509.363122] env[62385]: DEBUG nova.compute.manager [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 509.838194] env[62385]: DEBUG nova.compute.utils [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 509.845893] env[62385]: DEBUG nova.compute.manager [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 509.845893] env[62385]: DEBUG nova.network.neutron [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 509.895986] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 510.351278] env[62385]: DEBUG nova.compute.manager [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 510.493565] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7be2b1e-8386-4307-835c-73bda0cde64a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 510.506045] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb686887-1d2f-401b-accf-58e50f3672df {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 510.551826] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ffd7e42-e2fa-4ae2-be41-0d37501a16cc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 510.563254] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89cfd416-fa06-4d5a-af6d-52f624440fe3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 510.580823] env[62385]: DEBUG nova.compute.provider_tree [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 510.993025] env[62385]: DEBUG nova.policy [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '974da90b66294f06b79136241cc3cfd8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ad9b1ba20af94ec6ae8819c24bb3f5a4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 511.087783] env[62385]: DEBUG nova.scheduler.client.report [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 511.371679] env[62385]: DEBUG nova.compute.manager [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 511.410694] env[62385]: DEBUG nova.virt.hardware [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 511.410694] env[62385]: DEBUG nova.virt.hardware [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 511.410694] env[62385]: DEBUG nova.virt.hardware [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 511.410981] env[62385]: DEBUG nova.virt.hardware [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 511.411190] env[62385]: DEBUG nova.virt.hardware [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 511.411374] env[62385]: DEBUG nova.virt.hardware [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 511.411628] env[62385]: DEBUG nova.virt.hardware [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 511.411817] env[62385]: DEBUG nova.virt.hardware [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 511.412300] env[62385]: DEBUG nova.virt.hardware [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 511.412684] env[62385]: DEBUG nova.virt.hardware [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 511.413610] env[62385]: DEBUG nova.virt.hardware [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 511.414130] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-992d86aa-0386-4009-9c30-1fb7087f7c43 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.423892] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c7a173a-d3f5-46a4-a28f-a1b99fddf267 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.446400] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a796096-513a-4b22-8afd-1912c5ab9158 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.488743] env[62385]: DEBUG nova.network.neutron [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Successfully created port: dfae2567-f90c-4f41-a898-416ac4663fed {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 511.597548] env[62385]: DEBUG oslo_concurrency.lockutils [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.270s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 511.597548] env[62385]: DEBUG nova.compute.manager [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 511.600980] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.776s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 511.602353] env[62385]: INFO nova.compute.claims [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 512.113611] env[62385]: DEBUG nova.compute.utils [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 512.115208] env[62385]: DEBUG nova.compute.manager [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 512.115681] env[62385]: DEBUG nova.network.neutron [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 512.301965] env[62385]: DEBUG nova.policy [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '846f371a07c14ee899b9406244976167', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3409987c5d8742d384727cce2ffef5b1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 512.621169] env[62385]: DEBUG nova.compute.manager [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 512.765348] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff502eef-3ca4-4f7d-86ba-93bca0713ad6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.777497] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91117c17-0110-4021-9230-8e587d6c635a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.811975] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2a1bd96-05dd-4a14-8c9f-a38ddd386a15 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.824227] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ff54b8f-6614-456f-a827-25389aebb008 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.841849] env[62385]: DEBUG nova.compute.provider_tree [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 513.344985] env[62385]: DEBUG nova.scheduler.client.report [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 513.636848] env[62385]: DEBUG nova.compute.manager [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 513.666772] env[62385]: DEBUG nova.virt.hardware [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 513.667258] env[62385]: DEBUG nova.virt.hardware [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 513.667546] env[62385]: DEBUG nova.virt.hardware [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 513.667830] env[62385]: DEBUG nova.virt.hardware [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 513.668101] env[62385]: DEBUG nova.virt.hardware [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 513.668411] env[62385]: DEBUG nova.virt.hardware [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 513.670871] env[62385]: DEBUG nova.virt.hardware [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 513.670871] env[62385]: DEBUG nova.virt.hardware [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 513.670871] env[62385]: DEBUG nova.virt.hardware [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 513.670871] env[62385]: DEBUG nova.virt.hardware [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 513.670871] env[62385]: DEBUG nova.virt.hardware [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 513.671156] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5564483f-963a-4dfe-a04d-f27822388277 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.679633] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff8efc9d-98ab-4671-aba1-b87a1edb382e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.851494] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.250s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 513.853140] env[62385]: DEBUG nova.compute.manager [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 513.857772] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.308s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 513.859250] env[62385]: INFO nova.compute.claims [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 513.967511] env[62385]: DEBUG nova.network.neutron [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Successfully created port: fd5bfa71-824a-4d21-b913-3fbec4e93f31 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 514.371950] env[62385]: DEBUG nova.compute.utils [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 514.374077] env[62385]: DEBUG nova.compute.manager [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 514.374077] env[62385]: DEBUG nova.network.neutron [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 514.760009] env[62385]: DEBUG nova.policy [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7eca1fe46998449891939512039e8927', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '13922699f77642688ea51ac8ba0b421c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 514.878907] env[62385]: DEBUG nova.compute.manager [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 514.996155] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5297c56d-4fb3-4dc7-aea0-123e8a564fea {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.004371] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78a43cef-f657-4eed-9245-5fa64ae4a3a8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.042488] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8cc29ce-030f-4627-9930-1e644aa5aae9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.051715] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b86ddd8a-2fd6-4ecb-b6c4-a55d90b186cd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.071926] env[62385]: DEBUG nova.compute.provider_tree [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 515.146848] env[62385]: ERROR nova.compute.manager [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port dfae2567-f90c-4f41-a898-416ac4663fed, please check neutron logs for more information. [ 515.146848] env[62385]: ERROR nova.compute.manager Traceback (most recent call last): [ 515.146848] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 515.146848] env[62385]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 515.146848] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 515.146848] env[62385]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 515.146848] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 515.146848] env[62385]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 515.146848] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 515.146848] env[62385]: ERROR nova.compute.manager self.force_reraise() [ 515.146848] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 515.146848] env[62385]: ERROR nova.compute.manager raise self.value [ 515.146848] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 515.146848] env[62385]: ERROR nova.compute.manager updated_port = self._update_port( [ 515.146848] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 515.146848] env[62385]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 515.148029] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 515.148029] env[62385]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 515.148029] env[62385]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port dfae2567-f90c-4f41-a898-416ac4663fed, please check neutron logs for more information. [ 515.148029] env[62385]: ERROR nova.compute.manager [ 515.148390] env[62385]: Traceback (most recent call last): [ 515.148536] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 515.148536] env[62385]: listener.cb(fileno) [ 515.148536] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 515.148536] env[62385]: result = function(*args, **kwargs) [ 515.148536] env[62385]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 515.148536] env[62385]: return func(*args, **kwargs) [ 515.148536] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 515.148536] env[62385]: raise e [ 515.148536] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 515.148536] env[62385]: nwinfo = self.network_api.allocate_for_instance( [ 515.148979] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 515.148979] env[62385]: created_port_ids = self._update_ports_for_instance( [ 515.148979] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 515.148979] env[62385]: with excutils.save_and_reraise_exception(): [ 515.148979] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 515.148979] env[62385]: self.force_reraise() [ 515.148979] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 515.148979] env[62385]: raise self.value [ 515.148979] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 515.148979] env[62385]: updated_port = self._update_port( [ 515.148979] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 515.148979] env[62385]: _ensure_no_port_binding_failure(port) [ 515.148979] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 515.148979] env[62385]: raise exception.PortBindingFailed(port_id=port['id']) [ 515.148979] env[62385]: nova.exception.PortBindingFailed: Binding failed for port dfae2567-f90c-4f41-a898-416ac4663fed, please check neutron logs for more information. [ 515.148979] env[62385]: Removing descriptor: 16 [ 515.150550] env[62385]: ERROR nova.compute.manager [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port dfae2567-f90c-4f41-a898-416ac4663fed, please check neutron logs for more information. [ 515.150550] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Traceback (most recent call last): [ 515.150550] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 515.150550] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] yield resources [ 515.150550] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 515.150550] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] self.driver.spawn(context, instance, image_meta, [ 515.150550] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 515.150550] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 515.150550] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 515.150550] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] vm_ref = self.build_virtual_machine(instance, [ 515.150550] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 515.151037] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] vif_infos = vmwarevif.get_vif_info(self._session, [ 515.151037] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 515.151037] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] for vif in network_info: [ 515.151037] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 515.151037] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] return self._sync_wrapper(fn, *args, **kwargs) [ 515.151037] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 515.151037] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] self.wait() [ 515.151037] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 515.151037] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] self[:] = self._gt.wait() [ 515.151037] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 515.151037] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] return self._exit_event.wait() [ 515.151037] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 515.151037] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] result = hub.switch() [ 515.151353] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 515.151353] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] return self.greenlet.switch() [ 515.151353] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 515.151353] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] result = function(*args, **kwargs) [ 515.151353] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 515.151353] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] return func(*args, **kwargs) [ 515.151353] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 515.151353] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] raise e [ 515.151353] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 515.151353] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] nwinfo = self.network_api.allocate_for_instance( [ 515.151353] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 515.151353] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] created_port_ids = self._update_ports_for_instance( [ 515.151353] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 515.151664] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] with excutils.save_and_reraise_exception(): [ 515.151664] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 515.151664] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] self.force_reraise() [ 515.151664] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 515.151664] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] raise self.value [ 515.151664] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 515.151664] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] updated_port = self._update_port( [ 515.151664] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 515.151664] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] _ensure_no_port_binding_failure(port) [ 515.151664] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 515.151664] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] raise exception.PortBindingFailed(port_id=port['id']) [ 515.151664] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] nova.exception.PortBindingFailed: Binding failed for port dfae2567-f90c-4f41-a898-416ac4663fed, please check neutron logs for more information. [ 515.151664] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] [ 515.151981] env[62385]: INFO nova.compute.manager [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Terminating instance [ 515.155018] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] Acquiring lock "refresh_cache-ad6baae6-3925-41d7-9b95-0a5fe8add72d" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 515.155194] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] Acquired lock "refresh_cache-ad6baae6-3925-41d7-9b95-0a5fe8add72d" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 515.155410] env[62385]: DEBUG nova.network.neutron [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 515.576436] env[62385]: DEBUG nova.scheduler.client.report [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 515.719122] env[62385]: DEBUG nova.network.neutron [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 515.897952] env[62385]: DEBUG nova.compute.manager [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 515.947440] env[62385]: DEBUG nova.virt.hardware [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 515.947668] env[62385]: DEBUG nova.virt.hardware [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 515.947818] env[62385]: DEBUG nova.virt.hardware [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 515.947988] env[62385]: DEBUG nova.virt.hardware [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 515.948153] env[62385]: DEBUG nova.virt.hardware [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 515.948308] env[62385]: DEBUG nova.virt.hardware [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 515.948510] env[62385]: DEBUG nova.virt.hardware [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 515.948660] env[62385]: DEBUG nova.virt.hardware [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 515.948821] env[62385]: DEBUG nova.virt.hardware [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 515.948979] env[62385]: DEBUG nova.virt.hardware [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 515.949163] env[62385]: DEBUG nova.virt.hardware [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 515.950083] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5db2aed-a254-4524-9446-26ea4a26905b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.959714] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33d5459f-4763-4708-a082-f556d405a026 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.082892] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.225s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 516.083451] env[62385]: DEBUG nova.compute.manager [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 516.086257] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.191s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 516.087816] env[62385]: INFO nova.compute.claims [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 516.182292] env[62385]: DEBUG nova.network.neutron [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 516.439470] env[62385]: DEBUG oslo_concurrency.lockutils [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "ca158bd7-98e3-4ce0-9873-02571f3d9497" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 516.440298] env[62385]: DEBUG oslo_concurrency.lockutils [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "ca158bd7-98e3-4ce0-9873-02571f3d9497" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 516.579124] env[62385]: DEBUG nova.network.neutron [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Successfully created port: 06b7091d-aaf7-4853-8e5c-1569a815efa0 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 516.599748] env[62385]: DEBUG nova.compute.utils [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 516.605294] env[62385]: DEBUG nova.compute.manager [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 516.605294] env[62385]: DEBUG nova.network.neutron [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 516.690120] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] Releasing lock "refresh_cache-ad6baae6-3925-41d7-9b95-0a5fe8add72d" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 516.690531] env[62385]: DEBUG nova.compute.manager [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 516.690720] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 516.691502] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e6036ae3-061e-4fc1-95d9-d5a59bfcb747 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.702853] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90eb02c7-9e32-4cf5-a2c1-ed78f526c21f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.734535] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ad6baae6-3925-41d7-9b95-0a5fe8add72d could not be found. [ 516.736345] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 516.736403] env[62385]: INFO nova.compute.manager [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Took 0.05 seconds to destroy the instance on the hypervisor. [ 516.736727] env[62385]: DEBUG oslo.service.loopingcall [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 516.736985] env[62385]: DEBUG nova.compute.manager [-] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 516.737139] env[62385]: DEBUG nova.network.neutron [-] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 516.795531] env[62385]: DEBUG nova.network.neutron [-] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 516.848699] env[62385]: DEBUG nova.policy [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2f2a72e69a7749cfb5fe080e171cee47', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f7c6fcd602be4c998df22bca228c919d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 516.944060] env[62385]: DEBUG nova.compute.manager [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 517.043407] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 517.043749] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 517.043948] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Starting heal instance info cache {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 517.044152] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Rebuilding the list of instances to heal {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 517.104411] env[62385]: DEBUG nova.compute.manager [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 517.196673] env[62385]: DEBUG oslo_concurrency.lockutils [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] Acquiring lock "e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.196869] env[62385]: DEBUG oslo_concurrency.lockutils [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] Lock "e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 517.288892] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-485b6d3a-d347-4d07-9904-6e7fe13dfadc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.300244] env[62385]: DEBUG nova.network.neutron [-] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 517.304279] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16328224-83c4-48ab-a3af-429a97540fda {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.338491] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc4c9d63-9df7-45e5-af09-20b3d2c1518b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.346600] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5eacfb5-11e4-44cb-a68c-fac338bf8386 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.361297] env[62385]: DEBUG nova.compute.provider_tree [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 517.476046] env[62385]: DEBUG oslo_concurrency.lockutils [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.503581] env[62385]: ERROR nova.compute.manager [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fd5bfa71-824a-4d21-b913-3fbec4e93f31, please check neutron logs for more information. [ 517.503581] env[62385]: ERROR nova.compute.manager Traceback (most recent call last): [ 517.503581] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 517.503581] env[62385]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 517.503581] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 517.503581] env[62385]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 517.503581] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 517.503581] env[62385]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 517.503581] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 517.503581] env[62385]: ERROR nova.compute.manager self.force_reraise() [ 517.503581] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 517.503581] env[62385]: ERROR nova.compute.manager raise self.value [ 517.503581] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 517.503581] env[62385]: ERROR nova.compute.manager updated_port = self._update_port( [ 517.503581] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 517.503581] env[62385]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 517.504043] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 517.504043] env[62385]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 517.504043] env[62385]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fd5bfa71-824a-4d21-b913-3fbec4e93f31, please check neutron logs for more information. [ 517.504043] env[62385]: ERROR nova.compute.manager [ 517.504043] env[62385]: Traceback (most recent call last): [ 517.504043] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 517.504043] env[62385]: listener.cb(fileno) [ 517.504043] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 517.504043] env[62385]: result = function(*args, **kwargs) [ 517.504043] env[62385]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 517.504043] env[62385]: return func(*args, **kwargs) [ 517.504043] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 517.504043] env[62385]: raise e [ 517.504043] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 517.504043] env[62385]: nwinfo = self.network_api.allocate_for_instance( [ 517.504043] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 517.504043] env[62385]: created_port_ids = self._update_ports_for_instance( [ 517.504043] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 517.504043] env[62385]: with excutils.save_and_reraise_exception(): [ 517.504043] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 517.504043] env[62385]: self.force_reraise() [ 517.504043] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 517.504043] env[62385]: raise self.value [ 517.504043] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 517.504043] env[62385]: updated_port = self._update_port( [ 517.504043] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 517.504043] env[62385]: _ensure_no_port_binding_failure(port) [ 517.504043] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 517.504043] env[62385]: raise exception.PortBindingFailed(port_id=port['id']) [ 517.504765] env[62385]: nova.exception.PortBindingFailed: Binding failed for port fd5bfa71-824a-4d21-b913-3fbec4e93f31, please check neutron logs for more information. [ 517.504765] env[62385]: Removing descriptor: 17 [ 517.508158] env[62385]: ERROR nova.compute.manager [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fd5bfa71-824a-4d21-b913-3fbec4e93f31, please check neutron logs for more information. [ 517.508158] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Traceback (most recent call last): [ 517.508158] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 517.508158] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] yield resources [ 517.508158] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 517.508158] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] self.driver.spawn(context, instance, image_meta, [ 517.508158] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 517.508158] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 517.508158] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 517.508158] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] vm_ref = self.build_virtual_machine(instance, [ 517.508158] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 517.508593] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] vif_infos = vmwarevif.get_vif_info(self._session, [ 517.508593] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 517.508593] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] for vif in network_info: [ 517.508593] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 517.508593] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] return self._sync_wrapper(fn, *args, **kwargs) [ 517.508593] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 517.508593] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] self.wait() [ 517.508593] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 517.508593] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] self[:] = self._gt.wait() [ 517.508593] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 517.508593] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] return self._exit_event.wait() [ 517.508593] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 517.508593] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] result = hub.switch() [ 517.508927] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 517.508927] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] return self.greenlet.switch() [ 517.508927] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 517.508927] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] result = function(*args, **kwargs) [ 517.508927] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 517.508927] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] return func(*args, **kwargs) [ 517.508927] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 517.508927] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] raise e [ 517.508927] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 517.508927] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] nwinfo = self.network_api.allocate_for_instance( [ 517.508927] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 517.508927] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] created_port_ids = self._update_ports_for_instance( [ 517.508927] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 517.509323] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] with excutils.save_and_reraise_exception(): [ 517.509323] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 517.509323] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] self.force_reraise() [ 517.509323] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 517.509323] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] raise self.value [ 517.509323] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 517.509323] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] updated_port = self._update_port( [ 517.509323] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 517.509323] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] _ensure_no_port_binding_failure(port) [ 517.509323] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 517.509323] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] raise exception.PortBindingFailed(port_id=port['id']) [ 517.509323] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] nova.exception.PortBindingFailed: Binding failed for port fd5bfa71-824a-4d21-b913-3fbec4e93f31, please check neutron logs for more information. [ 517.509323] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] [ 517.509663] env[62385]: INFO nova.compute.manager [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Terminating instance [ 517.510435] env[62385]: DEBUG oslo_concurrency.lockutils [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] Acquiring lock "refresh_cache-adb14669-5f0b-4d3a-9cbf-293117a1d1cd" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 517.510591] env[62385]: DEBUG oslo_concurrency.lockutils [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] Acquired lock "refresh_cache-adb14669-5f0b-4d3a-9cbf-293117a1d1cd" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 517.510750] env[62385]: DEBUG nova.network.neutron [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 517.550875] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Skipping network cache update for instance because it is Building. {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 517.551154] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Skipping network cache update for instance because it is Building. {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 517.551217] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Skipping network cache update for instance because it is Building. {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 517.551343] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Skipping network cache update for instance because it is Building. {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 517.551486] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Skipping network cache update for instance because it is Building. {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 517.551604] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Didn't find any instances for network info cache update. {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 517.551794] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 517.552039] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 517.552233] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 517.552416] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 517.552596] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 517.552812] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 517.552961] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62385) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 517.553143] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager.update_available_resource {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 517.573480] env[62385]: DEBUG nova.compute.manager [req-feb5bd18-746a-4a63-9809-29a166fc126a req-b008f23b-1c9f-47fe-ad53-60ed3873e8b2 service nova] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Received event network-changed-dfae2567-f90c-4f41-a898-416ac4663fed {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 517.573681] env[62385]: DEBUG nova.compute.manager [req-feb5bd18-746a-4a63-9809-29a166fc126a req-b008f23b-1c9f-47fe-ad53-60ed3873e8b2 service nova] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Refreshing instance network info cache due to event network-changed-dfae2567-f90c-4f41-a898-416ac4663fed. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 517.573896] env[62385]: DEBUG oslo_concurrency.lockutils [req-feb5bd18-746a-4a63-9809-29a166fc126a req-b008f23b-1c9f-47fe-ad53-60ed3873e8b2 service nova] Acquiring lock "refresh_cache-ad6baae6-3925-41d7-9b95-0a5fe8add72d" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 517.574031] env[62385]: DEBUG oslo_concurrency.lockutils [req-feb5bd18-746a-4a63-9809-29a166fc126a req-b008f23b-1c9f-47fe-ad53-60ed3873e8b2 service nova] Acquired lock "refresh_cache-ad6baae6-3925-41d7-9b95-0a5fe8add72d" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 517.574186] env[62385]: DEBUG nova.network.neutron [req-feb5bd18-746a-4a63-9809-29a166fc126a req-b008f23b-1c9f-47fe-ad53-60ed3873e8b2 service nova] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Refreshing network info cache for port dfae2567-f90c-4f41-a898-416ac4663fed {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 517.701722] env[62385]: DEBUG nova.compute.manager [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 517.806701] env[62385]: INFO nova.compute.manager [-] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Took 1.07 seconds to deallocate network for instance. [ 517.812283] env[62385]: DEBUG nova.compute.claims [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Aborting claim: {{(pid=62385) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 517.812460] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.866212] env[62385]: DEBUG nova.scheduler.client.report [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 518.056260] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.062761] env[62385]: DEBUG nova.network.neutron [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 518.117545] env[62385]: DEBUG nova.compute.manager [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 518.147424] env[62385]: DEBUG nova.virt.hardware [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 518.148833] env[62385]: DEBUG nova.virt.hardware [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 518.148833] env[62385]: DEBUG nova.virt.hardware [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 518.148833] env[62385]: DEBUG nova.virt.hardware [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 518.148833] env[62385]: DEBUG nova.virt.hardware [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 518.148833] env[62385]: DEBUG nova.virt.hardware [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 518.149170] env[62385]: DEBUG nova.virt.hardware [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 518.149329] env[62385]: DEBUG nova.virt.hardware [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 518.149530] env[62385]: DEBUG nova.virt.hardware [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 518.149721] env[62385]: DEBUG nova.virt.hardware [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 518.149937] env[62385]: DEBUG nova.virt.hardware [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 518.151018] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6c28ef0-62d8-4936-a6d8-ba0df8d6884a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.155611] env[62385]: DEBUG nova.network.neutron [req-feb5bd18-746a-4a63-9809-29a166fc126a req-b008f23b-1c9f-47fe-ad53-60ed3873e8b2 service nova] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 518.162604] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b02faeae-a561-4a95-add6-7dacee8254ca {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.213385] env[62385]: DEBUG nova.network.neutron [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 518.232273] env[62385]: DEBUG oslo_concurrency.lockutils [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.372824] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.287s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 518.373413] env[62385]: DEBUG nova.compute.manager [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 518.377920] env[62385]: DEBUG oslo_concurrency.lockutils [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.901s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 518.378706] env[62385]: INFO nova.compute.claims [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 518.577051] env[62385]: DEBUG nova.network.neutron [req-feb5bd18-746a-4a63-9809-29a166fc126a req-b008f23b-1c9f-47fe-ad53-60ed3873e8b2 service nova] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 518.599291] env[62385]: DEBUG nova.network.neutron [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Successfully created port: 68a8e29c-58dc-4dbb-9f80-7dc9bbca6034 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 518.719203] env[62385]: DEBUG oslo_concurrency.lockutils [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] Releasing lock "refresh_cache-adb14669-5f0b-4d3a-9cbf-293117a1d1cd" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 518.719641] env[62385]: DEBUG nova.compute.manager [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 518.719840] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 518.720154] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e806958b-cefc-4a4b-b259-5032cf5de73a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.731827] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cb6732f-1f95-4380-b1e4-fca765aec249 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.755414] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance adb14669-5f0b-4d3a-9cbf-293117a1d1cd could not be found. [ 518.755918] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 518.755918] env[62385]: INFO nova.compute.manager [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Took 0.04 seconds to destroy the instance on the hypervisor. [ 518.756370] env[62385]: DEBUG oslo.service.loopingcall [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 518.756370] env[62385]: DEBUG nova.compute.manager [-] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 518.756482] env[62385]: DEBUG nova.network.neutron [-] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 518.764205] env[62385]: DEBUG nova.compute.manager [req-9fa5ad48-b3d0-4ca0-8ec4-7b0f6730d525 req-250649e5-48d4-415b-8738-0b1b3fed4044 service nova] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Received event network-changed-fd5bfa71-824a-4d21-b913-3fbec4e93f31 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 518.764205] env[62385]: DEBUG nova.compute.manager [req-9fa5ad48-b3d0-4ca0-8ec4-7b0f6730d525 req-250649e5-48d4-415b-8738-0b1b3fed4044 service nova] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Refreshing instance network info cache due to event network-changed-fd5bfa71-824a-4d21-b913-3fbec4e93f31. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 518.764205] env[62385]: DEBUG oslo_concurrency.lockutils [req-9fa5ad48-b3d0-4ca0-8ec4-7b0f6730d525 req-250649e5-48d4-415b-8738-0b1b3fed4044 service nova] Acquiring lock "refresh_cache-adb14669-5f0b-4d3a-9cbf-293117a1d1cd" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 518.764205] env[62385]: DEBUG oslo_concurrency.lockutils [req-9fa5ad48-b3d0-4ca0-8ec4-7b0f6730d525 req-250649e5-48d4-415b-8738-0b1b3fed4044 service nova] Acquired lock "refresh_cache-adb14669-5f0b-4d3a-9cbf-293117a1d1cd" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 518.764205] env[62385]: DEBUG nova.network.neutron [req-9fa5ad48-b3d0-4ca0-8ec4-7b0f6730d525 req-250649e5-48d4-415b-8738-0b1b3fed4044 service nova] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Refreshing network info cache for port fd5bfa71-824a-4d21-b913-3fbec4e93f31 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 518.803603] env[62385]: DEBUG nova.network.neutron [-] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 518.887617] env[62385]: DEBUG nova.compute.utils [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 518.888722] env[62385]: DEBUG nova.compute.manager [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 518.888884] env[62385]: DEBUG nova.network.neutron [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 518.980699] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] Acquiring lock "b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.981416] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] Lock "b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 519.080556] env[62385]: DEBUG oslo_concurrency.lockutils [req-feb5bd18-746a-4a63-9809-29a166fc126a req-b008f23b-1c9f-47fe-ad53-60ed3873e8b2 service nova] Releasing lock "refresh_cache-ad6baae6-3925-41d7-9b95-0a5fe8add72d" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 519.174038] env[62385]: DEBUG nova.policy [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9d0da5e5b9f54856ad914b313719b824', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '26ce94e5e21c4a64aebba9fb08918678', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 519.203430] env[62385]: DEBUG oslo_concurrency.lockutils [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] Acquiring lock "e075eece-534b-4d97-8c82-abb960ef1abe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.203430] env[62385]: DEBUG oslo_concurrency.lockutils [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] Lock "e075eece-534b-4d97-8c82-abb960ef1abe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 519.307411] env[62385]: DEBUG nova.network.neutron [-] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 519.394409] env[62385]: DEBUG nova.compute.manager [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 519.411737] env[62385]: DEBUG nova.network.neutron [req-9fa5ad48-b3d0-4ca0-8ec4-7b0f6730d525 req-250649e5-48d4-415b-8738-0b1b3fed4044 service nova] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 519.487523] env[62385]: DEBUG nova.compute.manager [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 519.512758] env[62385]: DEBUG nova.network.neutron [req-9fa5ad48-b3d0-4ca0-8ec4-7b0f6730d525 req-250649e5-48d4-415b-8738-0b1b3fed4044 service nova] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 519.581665] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2460082c-4cb6-476a-959b-7c5ba9096aa8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.590296] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76ea67c1-8686-4164-82e8-d3928ac976a5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.628939] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac96d094-ae10-423a-9bff-e3a81c94b04a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.638797] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2cb21d7-59ec-43c9-9405-90fc5cacfc39 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.656220] env[62385]: DEBUG nova.compute.provider_tree [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 519.707110] env[62385]: DEBUG nova.compute.manager [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 519.810167] env[62385]: INFO nova.compute.manager [-] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Took 1.05 seconds to deallocate network for instance. [ 519.814253] env[62385]: DEBUG nova.compute.claims [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Aborting claim: {{(pid=62385) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 519.814455] env[62385]: DEBUG oslo_concurrency.lockutils [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.016999] env[62385]: DEBUG oslo_concurrency.lockutils [req-9fa5ad48-b3d0-4ca0-8ec4-7b0f6730d525 req-250649e5-48d4-415b-8738-0b1b3fed4044 service nova] Releasing lock "refresh_cache-adb14669-5f0b-4d3a-9cbf-293117a1d1cd" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 520.027128] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.166203] env[62385]: DEBUG nova.scheduler.client.report [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 520.247423] env[62385]: DEBUG oslo_concurrency.lockutils [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.409993] env[62385]: DEBUG nova.compute.manager [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 520.437457] env[62385]: DEBUG nova.virt.hardware [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 520.437708] env[62385]: DEBUG nova.virt.hardware [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 520.437868] env[62385]: DEBUG nova.virt.hardware [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 520.438708] env[62385]: DEBUG nova.virt.hardware [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 520.439117] env[62385]: DEBUG nova.virt.hardware [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 520.439117] env[62385]: DEBUG nova.virt.hardware [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 520.439317] env[62385]: DEBUG nova.virt.hardware [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 520.439552] env[62385]: DEBUG nova.virt.hardware [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 520.439911] env[62385]: DEBUG nova.virt.hardware [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 520.439911] env[62385]: DEBUG nova.virt.hardware [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 520.440054] env[62385]: DEBUG nova.virt.hardware [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 520.441274] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66661827-de94-4755-9560-d44e33f65f4c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.451454] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2e2f744-ebf3-4af9-a269-7c9bdf867895 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.672394] env[62385]: DEBUG oslo_concurrency.lockutils [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.295s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 520.676138] env[62385]: DEBUG nova.compute.manager [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 520.676943] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 2.864s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 521.183278] env[62385]: DEBUG nova.compute.utils [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 521.190548] env[62385]: DEBUG nova.compute.manager [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 521.190548] env[62385]: DEBUG nova.network.neutron [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 521.248673] env[62385]: DEBUG nova.network.neutron [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Successfully created port: 9c5c2b66-3082-48a8-bec4-cc04e521703a {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 521.281628] env[62385]: DEBUG nova.policy [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e1200cdf75b04233b4c3f4626d4e27cf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aea48968d4464ac0ab01c998c059107d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 521.530757] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16ddb0b5-8ab8-48dd-9d74-9a3fce38a515 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.538851] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bc5a506-16db-4361-8990-eaca169d9552 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.576757] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aae6e44f-8acc-4334-81d9-1fe35dbba65b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.585515] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63c53888-67e6-477b-8dfd-e5f592300fc9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.599767] env[62385]: DEBUG nova.compute.provider_tree [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 521.690312] env[62385]: DEBUG nova.compute.manager [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 521.766372] env[62385]: DEBUG nova.network.neutron [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Successfully created port: 85746a66-c8a0-4b2b-8298-31705b9ef890 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 522.102697] env[62385]: DEBUG nova.scheduler.client.report [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 522.358657] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] Acquiring lock "15703806-9f67-424d-aadb-634634810993" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 522.358657] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] Lock "15703806-9f67-424d-aadb-634634810993" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 522.576402] env[62385]: ERROR nova.compute.manager [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 06b7091d-aaf7-4853-8e5c-1569a815efa0, please check neutron logs for more information. [ 522.576402] env[62385]: ERROR nova.compute.manager Traceback (most recent call last): [ 522.576402] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 522.576402] env[62385]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 522.576402] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 522.576402] env[62385]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 522.576402] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 522.576402] env[62385]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 522.576402] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 522.576402] env[62385]: ERROR nova.compute.manager self.force_reraise() [ 522.576402] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 522.576402] env[62385]: ERROR nova.compute.manager raise self.value [ 522.576402] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 522.576402] env[62385]: ERROR nova.compute.manager updated_port = self._update_port( [ 522.576402] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 522.576402] env[62385]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 522.576864] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 522.576864] env[62385]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 522.576864] env[62385]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 06b7091d-aaf7-4853-8e5c-1569a815efa0, please check neutron logs for more information. [ 522.576864] env[62385]: ERROR nova.compute.manager [ 522.576864] env[62385]: Traceback (most recent call last): [ 522.576864] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 522.576864] env[62385]: listener.cb(fileno) [ 522.576864] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 522.576864] env[62385]: result = function(*args, **kwargs) [ 522.576864] env[62385]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 522.576864] env[62385]: return func(*args, **kwargs) [ 522.576864] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 522.576864] env[62385]: raise e [ 522.576864] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 522.576864] env[62385]: nwinfo = self.network_api.allocate_for_instance( [ 522.576864] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 522.576864] env[62385]: created_port_ids = self._update_ports_for_instance( [ 522.576864] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 522.576864] env[62385]: with excutils.save_and_reraise_exception(): [ 522.576864] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 522.576864] env[62385]: self.force_reraise() [ 522.576864] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 522.576864] env[62385]: raise self.value [ 522.576864] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 522.576864] env[62385]: updated_port = self._update_port( [ 522.576864] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 522.576864] env[62385]: _ensure_no_port_binding_failure(port) [ 522.576864] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 522.576864] env[62385]: raise exception.PortBindingFailed(port_id=port['id']) [ 522.577588] env[62385]: nova.exception.PortBindingFailed: Binding failed for port 06b7091d-aaf7-4853-8e5c-1569a815efa0, please check neutron logs for more information. [ 522.577588] env[62385]: Removing descriptor: 18 [ 522.577588] env[62385]: ERROR nova.compute.manager [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 06b7091d-aaf7-4853-8e5c-1569a815efa0, please check neutron logs for more information. [ 522.577588] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Traceback (most recent call last): [ 522.577588] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 522.577588] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] yield resources [ 522.577588] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 522.577588] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] self.driver.spawn(context, instance, image_meta, [ 522.577588] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 522.577588] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] self._vmops.spawn(context, instance, image_meta, injected_files, [ 522.577588] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 522.577588] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] vm_ref = self.build_virtual_machine(instance, [ 522.577887] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 522.577887] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] vif_infos = vmwarevif.get_vif_info(self._session, [ 522.577887] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 522.577887] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] for vif in network_info: [ 522.577887] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 522.577887] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] return self._sync_wrapper(fn, *args, **kwargs) [ 522.577887] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 522.577887] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] self.wait() [ 522.577887] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 522.577887] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] self[:] = self._gt.wait() [ 522.577887] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 522.577887] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] return self._exit_event.wait() [ 522.577887] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 522.578210] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] result = hub.switch() [ 522.578210] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 522.578210] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] return self.greenlet.switch() [ 522.578210] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 522.578210] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] result = function(*args, **kwargs) [ 522.578210] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 522.578210] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] return func(*args, **kwargs) [ 522.578210] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 522.578210] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] raise e [ 522.578210] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 522.578210] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] nwinfo = self.network_api.allocate_for_instance( [ 522.578210] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 522.578210] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] created_port_ids = self._update_ports_for_instance( [ 522.578523] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 522.578523] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] with excutils.save_and_reraise_exception(): [ 522.578523] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 522.578523] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] self.force_reraise() [ 522.578523] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 522.578523] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] raise self.value [ 522.578523] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 522.578523] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] updated_port = self._update_port( [ 522.578523] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 522.578523] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] _ensure_no_port_binding_failure(port) [ 522.578523] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 522.578523] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] raise exception.PortBindingFailed(port_id=port['id']) [ 522.578817] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] nova.exception.PortBindingFailed: Binding failed for port 06b7091d-aaf7-4853-8e5c-1569a815efa0, please check neutron logs for more information. [ 522.578817] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] [ 522.578817] env[62385]: INFO nova.compute.manager [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Terminating instance [ 522.581496] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] Acquiring lock "refresh_cache-d4f4eb24-ea78-4ab1-9786-818ea7769342" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 522.581611] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] Acquired lock "refresh_cache-d4f4eb24-ea78-4ab1-9786-818ea7769342" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 522.581880] env[62385]: DEBUG nova.network.neutron [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 522.613242] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.936s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 522.615428] env[62385]: ERROR nova.compute.manager [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port dfae2567-f90c-4f41-a898-416ac4663fed, please check neutron logs for more information. [ 522.615428] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Traceback (most recent call last): [ 522.615428] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 522.615428] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] self.driver.spawn(context, instance, image_meta, [ 522.615428] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 522.615428] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 522.615428] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 522.615428] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] vm_ref = self.build_virtual_machine(instance, [ 522.615428] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 522.615428] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] vif_infos = vmwarevif.get_vif_info(self._session, [ 522.615428] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 522.615865] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] for vif in network_info: [ 522.615865] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 522.615865] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] return self._sync_wrapper(fn, *args, **kwargs) [ 522.615865] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 522.615865] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] self.wait() [ 522.615865] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 522.615865] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] self[:] = self._gt.wait() [ 522.615865] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 522.615865] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] return self._exit_event.wait() [ 522.615865] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 522.615865] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] result = hub.switch() [ 522.615865] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 522.615865] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] return self.greenlet.switch() [ 522.616244] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 522.616244] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] result = function(*args, **kwargs) [ 522.616244] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 522.616244] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] return func(*args, **kwargs) [ 522.616244] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 522.616244] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] raise e [ 522.616244] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 522.616244] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] nwinfo = self.network_api.allocate_for_instance( [ 522.616244] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 522.616244] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] created_port_ids = self._update_ports_for_instance( [ 522.616244] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 522.616244] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] with excutils.save_and_reraise_exception(): [ 522.616244] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 522.616567] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] self.force_reraise() [ 522.616567] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 522.616567] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] raise self.value [ 522.616567] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 522.616567] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] updated_port = self._update_port( [ 522.616567] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 522.616567] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] _ensure_no_port_binding_failure(port) [ 522.616567] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 522.616567] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] raise exception.PortBindingFailed(port_id=port['id']) [ 522.616567] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] nova.exception.PortBindingFailed: Binding failed for port dfae2567-f90c-4f41-a898-416ac4663fed, please check neutron logs for more information. [ 522.616567] env[62385]: ERROR nova.compute.manager [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] [ 522.616834] env[62385]: DEBUG nova.compute.utils [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Binding failed for port dfae2567-f90c-4f41-a898-416ac4663fed, please check neutron logs for more information. {{(pid=62385) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 522.619786] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 4.562s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 522.619786] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 522.619786] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62385) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 522.619786] env[62385]: DEBUG oslo_concurrency.lockutils [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.387s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 522.620323] env[62385]: INFO nova.compute.claims [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 522.629191] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-435895de-8594-4921-8b2c-a04de3ae058a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.642055] env[62385]: DEBUG nova.compute.manager [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Build of instance ad6baae6-3925-41d7-9b95-0a5fe8add72d was re-scheduled: Binding failed for port dfae2567-f90c-4f41-a898-416ac4663fed, please check neutron logs for more information. {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 522.642554] env[62385]: DEBUG nova.compute.manager [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Unplugging VIFs for instance {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 522.642780] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] Acquiring lock "refresh_cache-ad6baae6-3925-41d7-9b95-0a5fe8add72d" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 522.642927] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] Acquired lock "refresh_cache-ad6baae6-3925-41d7-9b95-0a5fe8add72d" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 522.643114] env[62385]: DEBUG nova.network.neutron [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 522.645338] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41dff7a9-532b-4627-83ec-379a099b4926 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.670790] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5dd7e64-add9-472e-8e91-69556173a01a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.685187] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcf2078e-ea48-47ae-9af8-bdc237ceb477 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.720070] env[62385]: DEBUG nova.compute.manager [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 522.722155] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181401MB free_disk=155GB free_vcpus=48 pci_devices=None {{(pid=62385) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 522.722503] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 522.757106] env[62385]: DEBUG nova.virt.hardware [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 522.757106] env[62385]: DEBUG nova.virt.hardware [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 522.757106] env[62385]: DEBUG nova.virt.hardware [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 522.757409] env[62385]: DEBUG nova.virt.hardware [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 522.757596] env[62385]: DEBUG nova.virt.hardware [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 522.757744] env[62385]: DEBUG nova.virt.hardware [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 522.757958] env[62385]: DEBUG nova.virt.hardware [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 522.758771] env[62385]: DEBUG nova.virt.hardware [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 522.758771] env[62385]: DEBUG nova.virt.hardware [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 522.758771] env[62385]: DEBUG nova.virt.hardware [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 522.758771] env[62385]: DEBUG nova.virt.hardware [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 522.759819] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78f00f4c-3859-4eb0-b380-95123a68539a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.769445] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77ba71f6-cfe4-40bb-a8b4-7f02e0618b8a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.864125] env[62385]: DEBUG nova.compute.manager [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] [instance: 15703806-9f67-424d-aadb-634634810993] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 523.135423] env[62385]: DEBUG nova.network.neutron [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 523.205688] env[62385]: DEBUG nova.network.neutron [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 523.386845] env[62385]: DEBUG nova.network.neutron [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 523.402282] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 523.487445] env[62385]: ERROR nova.compute.manager [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 85746a66-c8a0-4b2b-8298-31705b9ef890, please check neutron logs for more information. [ 523.487445] env[62385]: ERROR nova.compute.manager Traceback (most recent call last): [ 523.487445] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 523.487445] env[62385]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 523.487445] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 523.487445] env[62385]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 523.487445] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 523.487445] env[62385]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 523.487445] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.487445] env[62385]: ERROR nova.compute.manager self.force_reraise() [ 523.487445] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.487445] env[62385]: ERROR nova.compute.manager raise self.value [ 523.487445] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 523.487445] env[62385]: ERROR nova.compute.manager updated_port = self._update_port( [ 523.487445] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.487445] env[62385]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 523.487891] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.487891] env[62385]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 523.487891] env[62385]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 85746a66-c8a0-4b2b-8298-31705b9ef890, please check neutron logs for more information. [ 523.487891] env[62385]: ERROR nova.compute.manager [ 523.487891] env[62385]: Traceback (most recent call last): [ 523.487891] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 523.487891] env[62385]: listener.cb(fileno) [ 523.487891] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 523.487891] env[62385]: result = function(*args, **kwargs) [ 523.487891] env[62385]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 523.487891] env[62385]: return func(*args, **kwargs) [ 523.487891] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 523.487891] env[62385]: raise e [ 523.487891] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 523.487891] env[62385]: nwinfo = self.network_api.allocate_for_instance( [ 523.487891] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 523.487891] env[62385]: created_port_ids = self._update_ports_for_instance( [ 523.487891] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 523.487891] env[62385]: with excutils.save_and_reraise_exception(): [ 523.487891] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.487891] env[62385]: self.force_reraise() [ 523.487891] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.487891] env[62385]: raise self.value [ 523.487891] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 523.487891] env[62385]: updated_port = self._update_port( [ 523.487891] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.487891] env[62385]: _ensure_no_port_binding_failure(port) [ 523.487891] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.487891] env[62385]: raise exception.PortBindingFailed(port_id=port['id']) [ 523.489159] env[62385]: nova.exception.PortBindingFailed: Binding failed for port 85746a66-c8a0-4b2b-8298-31705b9ef890, please check neutron logs for more information. [ 523.489159] env[62385]: Removing descriptor: 17 [ 523.489159] env[62385]: ERROR nova.compute.manager [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 85746a66-c8a0-4b2b-8298-31705b9ef890, please check neutron logs for more information. [ 523.489159] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Traceback (most recent call last): [ 523.489159] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 523.489159] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] yield resources [ 523.489159] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 523.489159] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] self.driver.spawn(context, instance, image_meta, [ 523.489159] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 523.489159] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] self._vmops.spawn(context, instance, image_meta, injected_files, [ 523.489159] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 523.489159] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] vm_ref = self.build_virtual_machine(instance, [ 523.489479] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 523.489479] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] vif_infos = vmwarevif.get_vif_info(self._session, [ 523.489479] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 523.489479] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] for vif in network_info: [ 523.489479] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 523.489479] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] return self._sync_wrapper(fn, *args, **kwargs) [ 523.489479] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 523.489479] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] self.wait() [ 523.489479] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 523.489479] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] self[:] = self._gt.wait() [ 523.489479] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 523.489479] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] return self._exit_event.wait() [ 523.489479] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 523.490595] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] result = hub.switch() [ 523.490595] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 523.490595] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] return self.greenlet.switch() [ 523.490595] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 523.490595] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] result = function(*args, **kwargs) [ 523.490595] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 523.490595] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] return func(*args, **kwargs) [ 523.490595] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 523.490595] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] raise e [ 523.490595] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 523.490595] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] nwinfo = self.network_api.allocate_for_instance( [ 523.490595] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 523.490595] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] created_port_ids = self._update_ports_for_instance( [ 523.490931] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 523.490931] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] with excutils.save_and_reraise_exception(): [ 523.490931] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.490931] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] self.force_reraise() [ 523.490931] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.490931] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] raise self.value [ 523.490931] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 523.490931] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] updated_port = self._update_port( [ 523.490931] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.490931] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] _ensure_no_port_binding_failure(port) [ 523.490931] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.490931] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] raise exception.PortBindingFailed(port_id=port['id']) [ 523.491268] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] nova.exception.PortBindingFailed: Binding failed for port 85746a66-c8a0-4b2b-8298-31705b9ef890, please check neutron logs for more information. [ 523.491268] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] [ 523.491268] env[62385]: INFO nova.compute.manager [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Terminating instance [ 523.495710] env[62385]: DEBUG oslo_concurrency.lockutils [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "refresh_cache-ca158bd7-98e3-4ce0-9873-02571f3d9497" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 523.496047] env[62385]: DEBUG oslo_concurrency.lockutils [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquired lock "refresh_cache-ca158bd7-98e3-4ce0-9873-02571f3d9497" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 523.496551] env[62385]: DEBUG nova.network.neutron [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 523.574148] env[62385]: DEBUG nova.network.neutron [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 523.834331] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14e3306f-0810-4cc7-a200-1b3ec7abbf0c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.851147] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4070642-8373-443a-b45d-d72c6a4d00ce {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.884157] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15f44ac1-2d86-447d-80a8-aa0ea0e76942 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.892053] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] Releasing lock "refresh_cache-d4f4eb24-ea78-4ab1-9786-818ea7769342" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 523.892711] env[62385]: DEBUG nova.compute.manager [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 523.892905] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 523.893310] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-05ce975d-a1db-487b-95b3-90de08b8639b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.897240] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49f404a6-5a51-4479-9726-11411da28cba {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.917792] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6005434-d8ab-44c0-8a76-257745236033 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.940894] env[62385]: DEBUG nova.compute.provider_tree [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 523.948121] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d4f4eb24-ea78-4ab1-9786-818ea7769342 could not be found. [ 523.948121] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 523.948121] env[62385]: INFO nova.compute.manager [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Took 0.05 seconds to destroy the instance on the hypervisor. [ 523.948121] env[62385]: DEBUG oslo.service.loopingcall [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 523.948121] env[62385]: DEBUG nova.compute.manager [-] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 523.948121] env[62385]: DEBUG nova.network.neutron [-] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 524.000328] env[62385]: DEBUG nova.network.neutron [-] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 524.055728] env[62385]: DEBUG nova.network.neutron [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 524.077957] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] Releasing lock "refresh_cache-ad6baae6-3925-41d7-9b95-0a5fe8add72d" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 524.078216] env[62385]: DEBUG nova.compute.manager [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 524.078689] env[62385]: DEBUG nova.compute.manager [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 524.078689] env[62385]: DEBUG nova.network.neutron [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 524.145346] env[62385]: DEBUG nova.network.neutron [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 524.433437] env[62385]: ERROR nova.compute.manager [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 68a8e29c-58dc-4dbb-9f80-7dc9bbca6034, please check neutron logs for more information. [ 524.433437] env[62385]: ERROR nova.compute.manager Traceback (most recent call last): [ 524.433437] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 524.433437] env[62385]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 524.433437] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 524.433437] env[62385]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 524.433437] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 524.433437] env[62385]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 524.433437] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 524.433437] env[62385]: ERROR nova.compute.manager self.force_reraise() [ 524.433437] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 524.433437] env[62385]: ERROR nova.compute.manager raise self.value [ 524.433437] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 524.433437] env[62385]: ERROR nova.compute.manager updated_port = self._update_port( [ 524.433437] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 524.433437] env[62385]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 524.434184] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 524.434184] env[62385]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 524.434184] env[62385]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 68a8e29c-58dc-4dbb-9f80-7dc9bbca6034, please check neutron logs for more information. [ 524.434184] env[62385]: ERROR nova.compute.manager [ 524.434184] env[62385]: Traceback (most recent call last): [ 524.434184] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 524.434184] env[62385]: listener.cb(fileno) [ 524.434184] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 524.434184] env[62385]: result = function(*args, **kwargs) [ 524.434184] env[62385]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 524.434184] env[62385]: return func(*args, **kwargs) [ 524.434184] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 524.434184] env[62385]: raise e [ 524.434184] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 524.434184] env[62385]: nwinfo = self.network_api.allocate_for_instance( [ 524.434184] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 524.434184] env[62385]: created_port_ids = self._update_ports_for_instance( [ 524.434184] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 524.434184] env[62385]: with excutils.save_and_reraise_exception(): [ 524.434184] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 524.434184] env[62385]: self.force_reraise() [ 524.434184] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 524.434184] env[62385]: raise self.value [ 524.434184] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 524.434184] env[62385]: updated_port = self._update_port( [ 524.434184] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 524.434184] env[62385]: _ensure_no_port_binding_failure(port) [ 524.434184] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 524.434184] env[62385]: raise exception.PortBindingFailed(port_id=port['id']) [ 524.435140] env[62385]: nova.exception.PortBindingFailed: Binding failed for port 68a8e29c-58dc-4dbb-9f80-7dc9bbca6034, please check neutron logs for more information. [ 524.435140] env[62385]: Removing descriptor: 16 [ 524.435140] env[62385]: ERROR nova.compute.manager [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 68a8e29c-58dc-4dbb-9f80-7dc9bbca6034, please check neutron logs for more information. [ 524.435140] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Traceback (most recent call last): [ 524.435140] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 524.435140] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] yield resources [ 524.435140] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 524.435140] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] self.driver.spawn(context, instance, image_meta, [ 524.435140] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 524.435140] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 524.435140] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 524.435140] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] vm_ref = self.build_virtual_machine(instance, [ 524.435516] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 524.435516] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] vif_infos = vmwarevif.get_vif_info(self._session, [ 524.435516] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 524.435516] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] for vif in network_info: [ 524.435516] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 524.435516] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] return self._sync_wrapper(fn, *args, **kwargs) [ 524.435516] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 524.435516] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] self.wait() [ 524.435516] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 524.435516] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] self[:] = self._gt.wait() [ 524.435516] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 524.435516] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] return self._exit_event.wait() [ 524.435516] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 524.436764] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] result = hub.switch() [ 524.436764] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 524.436764] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] return self.greenlet.switch() [ 524.436764] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 524.436764] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] result = function(*args, **kwargs) [ 524.436764] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 524.436764] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] return func(*args, **kwargs) [ 524.436764] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 524.436764] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] raise e [ 524.436764] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 524.436764] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] nwinfo = self.network_api.allocate_for_instance( [ 524.436764] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 524.436764] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] created_port_ids = self._update_ports_for_instance( [ 524.437336] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 524.437336] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] with excutils.save_and_reraise_exception(): [ 524.437336] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 524.437336] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] self.force_reraise() [ 524.437336] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 524.437336] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] raise self.value [ 524.437336] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 524.437336] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] updated_port = self._update_port( [ 524.437336] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 524.437336] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] _ensure_no_port_binding_failure(port) [ 524.437336] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 524.437336] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] raise exception.PortBindingFailed(port_id=port['id']) [ 524.437763] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] nova.exception.PortBindingFailed: Binding failed for port 68a8e29c-58dc-4dbb-9f80-7dc9bbca6034, please check neutron logs for more information. [ 524.437763] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] [ 524.437763] env[62385]: INFO nova.compute.manager [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Terminating instance [ 524.437763] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Acquiring lock "refresh_cache-077b25b0-3586-4c56-81fa-526ff9c0ed1f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 524.437763] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Acquired lock "refresh_cache-077b25b0-3586-4c56-81fa-526ff9c0ed1f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 524.437763] env[62385]: DEBUG nova.network.neutron [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 524.449175] env[62385]: DEBUG nova.scheduler.client.report [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 524.501427] env[62385]: DEBUG nova.network.neutron [-] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 524.530147] env[62385]: DEBUG nova.network.neutron [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 524.572963] env[62385]: DEBUG nova.compute.manager [req-fa9507e5-28fe-4f97-80c1-62d642e752c3 req-7b54379a-fa44-4dbc-94f3-4ad1de10da2f service nova] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Received event network-vif-deleted-fd5bfa71-824a-4d21-b913-3fbec4e93f31 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 524.649671] env[62385]: DEBUG nova.network.neutron [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 524.959761] env[62385]: DEBUG oslo_concurrency.lockutils [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.340s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 524.959761] env[62385]: DEBUG nova.compute.manager [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 524.962013] env[62385]: DEBUG oslo_concurrency.lockutils [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 5.148s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 525.004724] env[62385]: INFO nova.compute.manager [-] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Took 1.06 seconds to deallocate network for instance. [ 525.007227] env[62385]: DEBUG nova.compute.claims [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Aborting claim: {{(pid=62385) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 525.007739] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.026203] env[62385]: DEBUG nova.network.neutron [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 525.034540] env[62385]: DEBUG oslo_concurrency.lockutils [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Releasing lock "refresh_cache-ca158bd7-98e3-4ce0-9873-02571f3d9497" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 525.034891] env[62385]: DEBUG nova.compute.manager [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 525.035116] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 525.035416] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-af0dccc3-4f2c-48ac-9ebb-4ea334f2a32c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.050412] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c02cd271-94c7-4b5c-9659-7115378e6a30 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.084095] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ca158bd7-98e3-4ce0-9873-02571f3d9497 could not be found. [ 525.084095] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 525.084251] env[62385]: INFO nova.compute.manager [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Took 0.05 seconds to destroy the instance on the hypervisor. [ 525.086634] env[62385]: DEBUG oslo.service.loopingcall [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 525.086634] env[62385]: DEBUG nova.compute.manager [-] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 525.086634] env[62385]: DEBUG nova.network.neutron [-] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 525.133939] env[62385]: DEBUG nova.network.neutron [-] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 525.152948] env[62385]: INFO nova.compute.manager [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] [instance: ad6baae6-3925-41d7-9b95-0a5fe8add72d] Took 1.07 seconds to deallocate network for instance. [ 525.292530] env[62385]: DEBUG nova.network.neutron [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 525.470953] env[62385]: DEBUG nova.compute.utils [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 525.478906] env[62385]: DEBUG nova.compute.manager [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 525.479109] env[62385]: DEBUG nova.network.neutron [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 525.637291] env[62385]: DEBUG nova.network.neutron [-] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 525.671140] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72bf6442-dbef-42e9-b4a4-ca9020bf7c3c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.681873] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddf71794-80a4-474a-9cda-ef396387bf9a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.723615] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3226913-dd45-4688-a615-7cf64ad4cfbe {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.732559] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0702ab7e-ed5e-4f2d-9fed-0fde1630a999 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.749044] env[62385]: DEBUG nova.compute.provider_tree [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 525.763829] env[62385]: DEBUG nova.policy [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5c8863bf352a44deafb45b93259eb26c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '52425be256114432a284c026d5d7abea', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 525.797303] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Releasing lock "refresh_cache-077b25b0-3586-4c56-81fa-526ff9c0ed1f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 525.797303] env[62385]: DEBUG nova.compute.manager [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 525.797519] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 525.799667] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c465fbb8-f455-469d-b5a5-a1f1d9dc79b9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.811295] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-563a5797-5b36-4f03-b835-d12acae9535d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.847021] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 077b25b0-3586-4c56-81fa-526ff9c0ed1f could not be found. [ 525.847021] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 525.847021] env[62385]: INFO nova.compute.manager [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Took 0.05 seconds to destroy the instance on the hypervisor. [ 525.847021] env[62385]: DEBUG oslo.service.loopingcall [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 525.847021] env[62385]: DEBUG nova.compute.manager [-] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 525.847021] env[62385]: DEBUG nova.network.neutron [-] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 525.933283] env[62385]: DEBUG nova.network.neutron [-] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 525.983954] env[62385]: DEBUG nova.compute.manager [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 526.139518] env[62385]: INFO nova.compute.manager [-] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Took 1.05 seconds to deallocate network for instance. [ 526.143374] env[62385]: DEBUG nova.compute.claims [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Aborting claim: {{(pid=62385) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 526.143600] env[62385]: DEBUG oslo_concurrency.lockutils [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 526.194342] env[62385]: INFO nova.scheduler.client.report [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] Deleted allocations for instance ad6baae6-3925-41d7-9b95-0a5fe8add72d [ 526.251775] env[62385]: DEBUG nova.scheduler.client.report [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 526.438629] env[62385]: DEBUG nova.network.neutron [-] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 526.501169] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Acquiring lock "7ab1b11a-cf7c-4f6c-aeb2-c79681136721" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 526.501169] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Lock "7ab1b11a-cf7c-4f6c-aeb2-c79681136721" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 526.702511] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1ee93fed-ac80-47fe-97cc-fedadc72f0c4 tempest-AttachInterfacesV270Test-495607412 tempest-AttachInterfacesV270Test-495607412-project-member] Lock "ad6baae6-3925-41d7-9b95-0a5fe8add72d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.630s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 526.744503] env[62385]: ERROR nova.compute.manager [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9c5c2b66-3082-48a8-bec4-cc04e521703a, please check neutron logs for more information. [ 526.744503] env[62385]: ERROR nova.compute.manager Traceback (most recent call last): [ 526.744503] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 526.744503] env[62385]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 526.744503] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 526.744503] env[62385]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 526.744503] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 526.744503] env[62385]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 526.744503] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 526.744503] env[62385]: ERROR nova.compute.manager self.force_reraise() [ 526.744503] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 526.744503] env[62385]: ERROR nova.compute.manager raise self.value [ 526.744503] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 526.744503] env[62385]: ERROR nova.compute.manager updated_port = self._update_port( [ 526.744503] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 526.744503] env[62385]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 526.745310] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 526.745310] env[62385]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 526.745310] env[62385]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9c5c2b66-3082-48a8-bec4-cc04e521703a, please check neutron logs for more information. [ 526.745310] env[62385]: ERROR nova.compute.manager [ 526.745310] env[62385]: Traceback (most recent call last): [ 526.745310] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 526.745310] env[62385]: listener.cb(fileno) [ 526.745310] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 526.745310] env[62385]: result = function(*args, **kwargs) [ 526.745310] env[62385]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 526.745310] env[62385]: return func(*args, **kwargs) [ 526.745310] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 526.745310] env[62385]: raise e [ 526.745310] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 526.745310] env[62385]: nwinfo = self.network_api.allocate_for_instance( [ 526.745310] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 526.745310] env[62385]: created_port_ids = self._update_ports_for_instance( [ 526.745310] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 526.745310] env[62385]: with excutils.save_and_reraise_exception(): [ 526.745310] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 526.745310] env[62385]: self.force_reraise() [ 526.745310] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 526.745310] env[62385]: raise self.value [ 526.745310] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 526.745310] env[62385]: updated_port = self._update_port( [ 526.745310] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 526.745310] env[62385]: _ensure_no_port_binding_failure(port) [ 526.745310] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 526.745310] env[62385]: raise exception.PortBindingFailed(port_id=port['id']) [ 526.746763] env[62385]: nova.exception.PortBindingFailed: Binding failed for port 9c5c2b66-3082-48a8-bec4-cc04e521703a, please check neutron logs for more information. [ 526.746763] env[62385]: Removing descriptor: 19 [ 526.746763] env[62385]: ERROR nova.compute.manager [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9c5c2b66-3082-48a8-bec4-cc04e521703a, please check neutron logs for more information. [ 526.746763] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Traceback (most recent call last): [ 526.746763] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 526.746763] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] yield resources [ 526.746763] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 526.746763] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] self.driver.spawn(context, instance, image_meta, [ 526.746763] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 526.746763] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] self._vmops.spawn(context, instance, image_meta, injected_files, [ 526.746763] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 526.746763] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] vm_ref = self.build_virtual_machine(instance, [ 526.747143] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 526.747143] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] vif_infos = vmwarevif.get_vif_info(self._session, [ 526.747143] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 526.747143] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] for vif in network_info: [ 526.747143] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 526.747143] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] return self._sync_wrapper(fn, *args, **kwargs) [ 526.747143] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 526.747143] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] self.wait() [ 526.747143] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 526.747143] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] self[:] = self._gt.wait() [ 526.747143] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 526.747143] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] return self._exit_event.wait() [ 526.747143] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 526.747510] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] result = hub.switch() [ 526.747510] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 526.747510] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] return self.greenlet.switch() [ 526.747510] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 526.747510] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] result = function(*args, **kwargs) [ 526.747510] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 526.747510] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] return func(*args, **kwargs) [ 526.747510] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 526.747510] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] raise e [ 526.747510] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 526.747510] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] nwinfo = self.network_api.allocate_for_instance( [ 526.747510] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 526.747510] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] created_port_ids = self._update_ports_for_instance( [ 526.748585] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 526.748585] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] with excutils.save_and_reraise_exception(): [ 526.748585] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 526.748585] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] self.force_reraise() [ 526.748585] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 526.748585] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] raise self.value [ 526.748585] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 526.748585] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] updated_port = self._update_port( [ 526.748585] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 526.748585] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] _ensure_no_port_binding_failure(port) [ 526.748585] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 526.748585] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] raise exception.PortBindingFailed(port_id=port['id']) [ 526.748977] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] nova.exception.PortBindingFailed: Binding failed for port 9c5c2b66-3082-48a8-bec4-cc04e521703a, please check neutron logs for more information. [ 526.748977] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] [ 526.748977] env[62385]: INFO nova.compute.manager [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Terminating instance [ 526.753424] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Acquiring lock "refresh_cache-b28550f5-4f8e-43ac-a020-c025586c6688" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 526.753762] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Acquired lock "refresh_cache-b28550f5-4f8e-43ac-a020-c025586c6688" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 526.754106] env[62385]: DEBUG nova.network.neutron [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 526.756780] env[62385]: DEBUG oslo_concurrency.lockutils [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.795s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 526.757456] env[62385]: ERROR nova.compute.manager [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fd5bfa71-824a-4d21-b913-3fbec4e93f31, please check neutron logs for more information. [ 526.757456] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Traceback (most recent call last): [ 526.757456] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 526.757456] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] self.driver.spawn(context, instance, image_meta, [ 526.757456] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 526.757456] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 526.757456] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 526.757456] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] vm_ref = self.build_virtual_machine(instance, [ 526.757456] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 526.757456] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] vif_infos = vmwarevif.get_vif_info(self._session, [ 526.757456] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 526.757796] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] for vif in network_info: [ 526.757796] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 526.757796] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] return self._sync_wrapper(fn, *args, **kwargs) [ 526.757796] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 526.757796] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] self.wait() [ 526.757796] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 526.757796] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] self[:] = self._gt.wait() [ 526.757796] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 526.757796] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] return self._exit_event.wait() [ 526.757796] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 526.757796] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] result = hub.switch() [ 526.757796] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 526.757796] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] return self.greenlet.switch() [ 526.758154] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 526.758154] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] result = function(*args, **kwargs) [ 526.758154] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 526.758154] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] return func(*args, **kwargs) [ 526.758154] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 526.758154] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] raise e [ 526.758154] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 526.758154] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] nwinfo = self.network_api.allocate_for_instance( [ 526.758154] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 526.758154] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] created_port_ids = self._update_ports_for_instance( [ 526.758154] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 526.758154] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] with excutils.save_and_reraise_exception(): [ 526.758154] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 526.759389] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] self.force_reraise() [ 526.759389] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 526.759389] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] raise self.value [ 526.759389] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 526.759389] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] updated_port = self._update_port( [ 526.759389] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 526.759389] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] _ensure_no_port_binding_failure(port) [ 526.759389] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 526.759389] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] raise exception.PortBindingFailed(port_id=port['id']) [ 526.759389] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] nova.exception.PortBindingFailed: Binding failed for port fd5bfa71-824a-4d21-b913-3fbec4e93f31, please check neutron logs for more information. [ 526.759389] env[62385]: ERROR nova.compute.manager [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] [ 526.760474] env[62385]: DEBUG nova.compute.utils [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Binding failed for port fd5bfa71-824a-4d21-b913-3fbec4e93f31, please check neutron logs for more information. {{(pid=62385) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 526.760474] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.733s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 526.760899] env[62385]: INFO nova.compute.claims [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 526.763410] env[62385]: DEBUG nova.compute.manager [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Build of instance adb14669-5f0b-4d3a-9cbf-293117a1d1cd was re-scheduled: Binding failed for port fd5bfa71-824a-4d21-b913-3fbec4e93f31, please check neutron logs for more information. {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 526.764149] env[62385]: DEBUG nova.compute.manager [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Unplugging VIFs for instance {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 526.764149] env[62385]: DEBUG oslo_concurrency.lockutils [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] Acquiring lock "refresh_cache-adb14669-5f0b-4d3a-9cbf-293117a1d1cd" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 526.764149] env[62385]: DEBUG oslo_concurrency.lockutils [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] Acquired lock "refresh_cache-adb14669-5f0b-4d3a-9cbf-293117a1d1cd" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 526.764296] env[62385]: DEBUG nova.network.neutron [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 526.766988] env[62385]: DEBUG nova.compute.manager [req-5a722666-06b6-49ba-a75a-82b0e5336586 req-2b57e38b-0b8c-4586-bf9e-f80c753092b7 service nova] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Received event network-changed-06b7091d-aaf7-4853-8e5c-1569a815efa0 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 526.767180] env[62385]: DEBUG nova.compute.manager [req-5a722666-06b6-49ba-a75a-82b0e5336586 req-2b57e38b-0b8c-4586-bf9e-f80c753092b7 service nova] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Refreshing instance network info cache due to event network-changed-06b7091d-aaf7-4853-8e5c-1569a815efa0. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 526.767385] env[62385]: DEBUG oslo_concurrency.lockutils [req-5a722666-06b6-49ba-a75a-82b0e5336586 req-2b57e38b-0b8c-4586-bf9e-f80c753092b7 service nova] Acquiring lock "refresh_cache-d4f4eb24-ea78-4ab1-9786-818ea7769342" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 526.767506] env[62385]: DEBUG oslo_concurrency.lockutils [req-5a722666-06b6-49ba-a75a-82b0e5336586 req-2b57e38b-0b8c-4586-bf9e-f80c753092b7 service nova] Acquired lock "refresh_cache-d4f4eb24-ea78-4ab1-9786-818ea7769342" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 526.767654] env[62385]: DEBUG nova.network.neutron [req-5a722666-06b6-49ba-a75a-82b0e5336586 req-2b57e38b-0b8c-4586-bf9e-f80c753092b7 service nova] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Refreshing network info cache for port 06b7091d-aaf7-4853-8e5c-1569a815efa0 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 526.945050] env[62385]: INFO nova.compute.manager [-] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Took 1.10 seconds to deallocate network for instance. [ 526.946708] env[62385]: DEBUG nova.compute.claims [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Aborting claim: {{(pid=62385) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 526.946708] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 526.999679] env[62385]: DEBUG nova.compute.manager [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 527.030324] env[62385]: DEBUG nova.virt.hardware [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 527.030813] env[62385]: DEBUG nova.virt.hardware [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 527.030813] env[62385]: DEBUG nova.virt.hardware [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 527.030929] env[62385]: DEBUG nova.virt.hardware [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 527.031980] env[62385]: DEBUG nova.virt.hardware [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 527.032465] env[62385]: DEBUG nova.virt.hardware [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 527.032756] env[62385]: DEBUG nova.virt.hardware [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 527.033177] env[62385]: DEBUG nova.virt.hardware [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 527.033177] env[62385]: DEBUG nova.virt.hardware [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 527.033494] env[62385]: DEBUG nova.virt.hardware [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 527.033708] env[62385]: DEBUG nova.virt.hardware [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 527.035951] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-812af828-cfba-485d-8f8b-6c29cdcec9b5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.047244] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f734357-4721-465b-8485-58beaaac39c2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.128857] env[62385]: DEBUG nova.network.neutron [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Successfully created port: 22d16c35-10f1-49bd-b538-c969c7034480 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 527.208473] env[62385]: DEBUG nova.compute.manager [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 527.306669] env[62385]: DEBUG nova.network.neutron [req-5a722666-06b6-49ba-a75a-82b0e5336586 req-2b57e38b-0b8c-4586-bf9e-f80c753092b7 service nova] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 527.344589] env[62385]: DEBUG nova.network.neutron [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 527.348849] env[62385]: DEBUG nova.network.neutron [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 527.431432] env[62385]: DEBUG nova.network.neutron [req-5a722666-06b6-49ba-a75a-82b0e5336586 req-2b57e38b-0b8c-4586-bf9e-f80c753092b7 service nova] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 527.686628] env[62385]: DEBUG nova.network.neutron [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 527.699775] env[62385]: DEBUG nova.network.neutron [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 527.742398] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 527.942031] env[62385]: DEBUG oslo_concurrency.lockutils [req-5a722666-06b6-49ba-a75a-82b0e5336586 req-2b57e38b-0b8c-4586-bf9e-f80c753092b7 service nova] Releasing lock "refresh_cache-d4f4eb24-ea78-4ab1-9786-818ea7769342" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 527.942031] env[62385]: DEBUG nova.compute.manager [req-5a722666-06b6-49ba-a75a-82b0e5336586 req-2b57e38b-0b8c-4586-bf9e-f80c753092b7 service nova] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Received event network-vif-deleted-06b7091d-aaf7-4853-8e5c-1569a815efa0 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 527.945016] env[62385]: DEBUG nova.compute.manager [req-5a722666-06b6-49ba-a75a-82b0e5336586 req-2b57e38b-0b8c-4586-bf9e-f80c753092b7 service nova] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Received event network-changed-68a8e29c-58dc-4dbb-9f80-7dc9bbca6034 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 527.945016] env[62385]: DEBUG nova.compute.manager [req-5a722666-06b6-49ba-a75a-82b0e5336586 req-2b57e38b-0b8c-4586-bf9e-f80c753092b7 service nova] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Refreshing instance network info cache due to event network-changed-68a8e29c-58dc-4dbb-9f80-7dc9bbca6034. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 527.945016] env[62385]: DEBUG oslo_concurrency.lockutils [req-5a722666-06b6-49ba-a75a-82b0e5336586 req-2b57e38b-0b8c-4586-bf9e-f80c753092b7 service nova] Acquiring lock "refresh_cache-077b25b0-3586-4c56-81fa-526ff9c0ed1f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 527.945016] env[62385]: DEBUG oslo_concurrency.lockutils [req-5a722666-06b6-49ba-a75a-82b0e5336586 req-2b57e38b-0b8c-4586-bf9e-f80c753092b7 service nova] Acquired lock "refresh_cache-077b25b0-3586-4c56-81fa-526ff9c0ed1f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 527.945016] env[62385]: DEBUG nova.network.neutron [req-5a722666-06b6-49ba-a75a-82b0e5336586 req-2b57e38b-0b8c-4586-bf9e-f80c753092b7 service nova] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Refreshing network info cache for port 68a8e29c-58dc-4dbb-9f80-7dc9bbca6034 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 527.989424] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ace6d33c-9e82-4bf8-8278-1283fa2ec8e2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.998518] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96008986-cf3e-4330-bb2a-ff65f1eb0517 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.034986] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4bcb8f1-9fd5-4d0a-bb11-023744679bdc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.048390] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-784fb98e-4558-4f31-9add-4cb0ebce83c1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.066705] env[62385]: DEBUG nova.compute.provider_tree [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 528.189464] env[62385]: DEBUG oslo_concurrency.lockutils [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] Releasing lock "refresh_cache-adb14669-5f0b-4d3a-9cbf-293117a1d1cd" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 528.189800] env[62385]: DEBUG nova.compute.manager [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 528.190033] env[62385]: DEBUG nova.compute.manager [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 528.190446] env[62385]: DEBUG nova.network.neutron [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 528.203176] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Releasing lock "refresh_cache-b28550f5-4f8e-43ac-a020-c025586c6688" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 528.203176] env[62385]: DEBUG nova.compute.manager [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 528.203279] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 528.203567] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c15bf62b-e981-4b85-a0b5-9d759e4add85 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.215031] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ea9fb19-67cb-41dc-a968-bcf51e6dc954 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.230296] env[62385]: DEBUG nova.network.neutron [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 528.250037] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b28550f5-4f8e-43ac-a020-c025586c6688 could not be found. [ 528.250256] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 528.250448] env[62385]: INFO nova.compute.manager [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Took 0.05 seconds to destroy the instance on the hypervisor. [ 528.251323] env[62385]: DEBUG oslo.service.loopingcall [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 528.252022] env[62385]: DEBUG nova.compute.manager [-] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 528.252120] env[62385]: DEBUG nova.network.neutron [-] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 528.266721] env[62385]: DEBUG nova.compute.manager [req-31f85906-b122-4358-a47c-0f19f2652f4a req-25217742-afc3-4719-9e0a-9820a5a1b402 service nova] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Received event network-changed-85746a66-c8a0-4b2b-8298-31705b9ef890 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 528.266721] env[62385]: DEBUG nova.compute.manager [req-31f85906-b122-4358-a47c-0f19f2652f4a req-25217742-afc3-4719-9e0a-9820a5a1b402 service nova] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Refreshing instance network info cache due to event network-changed-85746a66-c8a0-4b2b-8298-31705b9ef890. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 528.266721] env[62385]: DEBUG oslo_concurrency.lockutils [req-31f85906-b122-4358-a47c-0f19f2652f4a req-25217742-afc3-4719-9e0a-9820a5a1b402 service nova] Acquiring lock "refresh_cache-ca158bd7-98e3-4ce0-9873-02571f3d9497" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 528.267021] env[62385]: DEBUG oslo_concurrency.lockutils [req-31f85906-b122-4358-a47c-0f19f2652f4a req-25217742-afc3-4719-9e0a-9820a5a1b402 service nova] Acquired lock "refresh_cache-ca158bd7-98e3-4ce0-9873-02571f3d9497" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 528.267021] env[62385]: DEBUG nova.network.neutron [req-31f85906-b122-4358-a47c-0f19f2652f4a req-25217742-afc3-4719-9e0a-9820a5a1b402 service nova] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Refreshing network info cache for port 85746a66-c8a0-4b2b-8298-31705b9ef890 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 528.307535] env[62385]: DEBUG nova.network.neutron [-] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 528.484754] env[62385]: DEBUG nova.network.neutron [req-5a722666-06b6-49ba-a75a-82b0e5336586 req-2b57e38b-0b8c-4586-bf9e-f80c753092b7 service nova] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 528.571541] env[62385]: DEBUG nova.scheduler.client.report [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 528.712880] env[62385]: DEBUG nova.network.neutron [req-5a722666-06b6-49ba-a75a-82b0e5336586 req-2b57e38b-0b8c-4586-bf9e-f80c753092b7 service nova] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 528.733468] env[62385]: DEBUG nova.network.neutron [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 528.810304] env[62385]: DEBUG nova.network.neutron [-] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 528.822136] env[62385]: DEBUG nova.network.neutron [req-31f85906-b122-4358-a47c-0f19f2652f4a req-25217742-afc3-4719-9e0a-9820a5a1b402 service nova] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 529.070155] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] Acquiring lock "d0f3ba00-5eee-4ff6-8999-3a318c9f0121" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 529.070377] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] Lock "d0f3ba00-5eee-4ff6-8999-3a318c9f0121" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 529.077443] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.318s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 529.077878] env[62385]: DEBUG nova.compute.manager [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 529.084442] env[62385]: DEBUG oslo_concurrency.lockutils [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.834s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 529.084442] env[62385]: INFO nova.compute.claims [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 529.125219] env[62385]: DEBUG nova.network.neutron [req-31f85906-b122-4358-a47c-0f19f2652f4a req-25217742-afc3-4719-9e0a-9820a5a1b402 service nova] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 529.216430] env[62385]: DEBUG oslo_concurrency.lockutils [req-5a722666-06b6-49ba-a75a-82b0e5336586 req-2b57e38b-0b8c-4586-bf9e-f80c753092b7 service nova] Releasing lock "refresh_cache-077b25b0-3586-4c56-81fa-526ff9c0ed1f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 529.216716] env[62385]: DEBUG nova.compute.manager [req-5a722666-06b6-49ba-a75a-82b0e5336586 req-2b57e38b-0b8c-4586-bf9e-f80c753092b7 service nova] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Received event network-vif-deleted-68a8e29c-58dc-4dbb-9f80-7dc9bbca6034 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 529.236774] env[62385]: INFO nova.compute.manager [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] [instance: adb14669-5f0b-4d3a-9cbf-293117a1d1cd] Took 1.05 seconds to deallocate network for instance. [ 529.315644] env[62385]: INFO nova.compute.manager [-] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Took 1.06 seconds to deallocate network for instance. [ 529.317682] env[62385]: DEBUG nova.compute.claims [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Aborting claim: {{(pid=62385) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 529.317852] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 529.583880] env[62385]: DEBUG nova.compute.utils [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 529.586160] env[62385]: DEBUG nova.compute.manager [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 529.586335] env[62385]: DEBUG nova.network.neutron [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 529.628856] env[62385]: DEBUG oslo_concurrency.lockutils [req-31f85906-b122-4358-a47c-0f19f2652f4a req-25217742-afc3-4719-9e0a-9820a5a1b402 service nova] Releasing lock "refresh_cache-ca158bd7-98e3-4ce0-9873-02571f3d9497" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 529.629116] env[62385]: DEBUG nova.compute.manager [req-31f85906-b122-4358-a47c-0f19f2652f4a req-25217742-afc3-4719-9e0a-9820a5a1b402 service nova] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Received event network-vif-deleted-85746a66-c8a0-4b2b-8298-31705b9ef890 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 529.748758] env[62385]: DEBUG nova.policy [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9cd532e0e12b4b5b8d939a3ce388146c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4b1d85ab72464eb08fae1909daddb452', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 530.093211] env[62385]: DEBUG nova.compute.manager [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 530.194305] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Acquiring lock "874d90b0-5657-467c-a934-31db63982ded" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 530.195519] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Lock "874d90b0-5657-467c-a934-31db63982ded" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 530.217746] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Acquiring lock "2abff399-f03d-4c62-ac0b-440ba54bec29" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 530.218132] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Lock "2abff399-f03d-4c62-ac0b-440ba54bec29" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 530.282171] env[62385]: INFO nova.scheduler.client.report [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] Deleted allocations for instance adb14669-5f0b-4d3a-9cbf-293117a1d1cd [ 530.356158] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c65886c7-0966-4f8f-9d23-788bcef818d6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.375355] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66f1b13b-76e7-4742-8c42-e0e6624dac08 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.432674] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c425167f-5887-4c3d-a436-cbd0f89ac2d4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.442025] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17e70b88-a418-4992-b4c1-e1c8a535e659 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.459462] env[62385]: DEBUG nova.compute.provider_tree [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 530.793219] env[62385]: DEBUG oslo_concurrency.lockutils [None req-37736a9e-1613-4235-8131-116dd0a2d828 tempest-TenantUsagesTestJSON-1683292940 tempest-TenantUsagesTestJSON-1683292940-project-member] Lock "adb14669-5f0b-4d3a-9cbf-293117a1d1cd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.729s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 530.944751] env[62385]: DEBUG nova.network.neutron [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Successfully created port: ed7ea7f6-f96a-40c0-86f0-88674a2954a4 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 530.964427] env[62385]: ERROR nova.compute.manager [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 22d16c35-10f1-49bd-b538-c969c7034480, please check neutron logs for more information. [ 530.964427] env[62385]: ERROR nova.compute.manager Traceback (most recent call last): [ 530.964427] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 530.964427] env[62385]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 530.964427] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 530.964427] env[62385]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 530.964427] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 530.964427] env[62385]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 530.964427] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 530.964427] env[62385]: ERROR nova.compute.manager self.force_reraise() [ 530.964427] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 530.964427] env[62385]: ERROR nova.compute.manager raise self.value [ 530.964427] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 530.964427] env[62385]: ERROR nova.compute.manager updated_port = self._update_port( [ 530.964427] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 530.964427] env[62385]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 530.964985] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 530.964985] env[62385]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 530.964985] env[62385]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 22d16c35-10f1-49bd-b538-c969c7034480, please check neutron logs for more information. [ 530.964985] env[62385]: ERROR nova.compute.manager [ 530.964985] env[62385]: Traceback (most recent call last): [ 530.964985] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 530.964985] env[62385]: listener.cb(fileno) [ 530.964985] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 530.964985] env[62385]: result = function(*args, **kwargs) [ 530.964985] env[62385]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 530.964985] env[62385]: return func(*args, **kwargs) [ 530.964985] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 530.964985] env[62385]: raise e [ 530.964985] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 530.964985] env[62385]: nwinfo = self.network_api.allocate_for_instance( [ 530.964985] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 530.964985] env[62385]: created_port_ids = self._update_ports_for_instance( [ 530.964985] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 530.964985] env[62385]: with excutils.save_and_reraise_exception(): [ 530.964985] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 530.964985] env[62385]: self.force_reraise() [ 530.964985] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 530.964985] env[62385]: raise self.value [ 530.964985] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 530.964985] env[62385]: updated_port = self._update_port( [ 530.964985] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 530.964985] env[62385]: _ensure_no_port_binding_failure(port) [ 530.964985] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 530.964985] env[62385]: raise exception.PortBindingFailed(port_id=port['id']) [ 530.965912] env[62385]: nova.exception.PortBindingFailed: Binding failed for port 22d16c35-10f1-49bd-b538-c969c7034480, please check neutron logs for more information. [ 530.965912] env[62385]: Removing descriptor: 20 [ 530.965912] env[62385]: DEBUG nova.scheduler.client.report [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 530.973038] env[62385]: ERROR nova.compute.manager [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 22d16c35-10f1-49bd-b538-c969c7034480, please check neutron logs for more information. [ 530.973038] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Traceback (most recent call last): [ 530.973038] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 530.973038] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] yield resources [ 530.973038] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 530.973038] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] self.driver.spawn(context, instance, image_meta, [ 530.973038] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 530.973038] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] self._vmops.spawn(context, instance, image_meta, injected_files, [ 530.973038] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 530.973038] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] vm_ref = self.build_virtual_machine(instance, [ 530.973038] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 530.973444] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] vif_infos = vmwarevif.get_vif_info(self._session, [ 530.973444] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 530.973444] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] for vif in network_info: [ 530.973444] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 530.973444] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] return self._sync_wrapper(fn, *args, **kwargs) [ 530.973444] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 530.973444] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] self.wait() [ 530.973444] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 530.973444] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] self[:] = self._gt.wait() [ 530.973444] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 530.973444] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] return self._exit_event.wait() [ 530.973444] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 530.973444] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] result = hub.switch() [ 530.973830] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 530.973830] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] return self.greenlet.switch() [ 530.973830] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 530.973830] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] result = function(*args, **kwargs) [ 530.973830] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 530.973830] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] return func(*args, **kwargs) [ 530.973830] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 530.973830] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] raise e [ 530.973830] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 530.973830] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] nwinfo = self.network_api.allocate_for_instance( [ 530.973830] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 530.973830] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] created_port_ids = self._update_ports_for_instance( [ 530.973830] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 530.975048] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] with excutils.save_and_reraise_exception(): [ 530.975048] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 530.975048] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] self.force_reraise() [ 530.975048] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 530.975048] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] raise self.value [ 530.975048] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 530.975048] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] updated_port = self._update_port( [ 530.975048] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 530.975048] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] _ensure_no_port_binding_failure(port) [ 530.975048] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 530.975048] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] raise exception.PortBindingFailed(port_id=port['id']) [ 530.975048] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] nova.exception.PortBindingFailed: Binding failed for port 22d16c35-10f1-49bd-b538-c969c7034480, please check neutron logs for more information. [ 530.975048] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] [ 530.975464] env[62385]: INFO nova.compute.manager [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Terminating instance [ 530.975464] env[62385]: DEBUG oslo_concurrency.lockutils [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] Acquiring lock "refresh_cache-e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 530.975464] env[62385]: DEBUG oslo_concurrency.lockutils [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] Acquired lock "refresh_cache-e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 530.975464] env[62385]: DEBUG nova.network.neutron [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 531.110228] env[62385]: DEBUG nova.compute.manager [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 531.137567] env[62385]: DEBUG nova.virt.hardware [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 531.137808] env[62385]: DEBUG nova.virt.hardware [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 531.137965] env[62385]: DEBUG nova.virt.hardware [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 531.138160] env[62385]: DEBUG nova.virt.hardware [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 531.138345] env[62385]: DEBUG nova.virt.hardware [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 531.138516] env[62385]: DEBUG nova.virt.hardware [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 531.138726] env[62385]: DEBUG nova.virt.hardware [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 531.140493] env[62385]: DEBUG nova.virt.hardware [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 531.140841] env[62385]: DEBUG nova.virt.hardware [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 531.141047] env[62385]: DEBUG nova.virt.hardware [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 531.144015] env[62385]: DEBUG nova.virt.hardware [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 531.144015] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-567882a2-03c0-4522-bec6-0f2dd3ba9a75 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.152298] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7a23018-ad44-43f8-956a-7e9e6de9263c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.299837] env[62385]: DEBUG nova.compute.manager [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 531.473566] env[62385]: DEBUG oslo_concurrency.lockutils [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.392s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 531.474156] env[62385]: DEBUG nova.compute.manager [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 531.478070] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 8.756s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.511640] env[62385]: DEBUG nova.network.neutron [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 531.745264] env[62385]: DEBUG nova.network.neutron [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 531.829797] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.930851] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Acquiring lock "9ddab157-12cf-4225-bc6f-e03ef9455d79" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.931172] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Lock "9ddab157-12cf-4225-bc6f-e03ef9455d79" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.991205] env[62385]: DEBUG nova.compute.utils [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 531.995505] env[62385]: DEBUG nova.compute.manager [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 531.999580] env[62385]: DEBUG nova.network.neutron [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 532.150968] env[62385]: DEBUG nova.policy [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ef9e3e55fd5d46a49afd247e7fb4f3db', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5433efa61cab42d2b9b012acc01ca237', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 532.248407] env[62385]: DEBUG oslo_concurrency.lockutils [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] Releasing lock "refresh_cache-e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 532.249587] env[62385]: DEBUG nova.compute.manager [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 532.249666] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 532.249941] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-68d58f1b-5b73-444f-8669-dab99bb30e31 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.261945] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10158f22-8a31-49e2-b92e-2441948e83ea {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.289370] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca could not be found. [ 532.289611] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 532.289932] env[62385]: INFO nova.compute.manager [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Took 0.04 seconds to destroy the instance on the hypervisor. [ 532.290057] env[62385]: DEBUG oslo.service.loopingcall [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 532.290279] env[62385]: DEBUG nova.compute.manager [-] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 532.290375] env[62385]: DEBUG nova.network.neutron [-] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 532.332024] env[62385]: DEBUG nova.network.neutron [-] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 532.502405] env[62385]: DEBUG nova.compute.manager [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 532.530500] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance d4f4eb24-ea78-4ab1-9786-818ea7769342 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 532.531930] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 077b25b0-3586-4c56-81fa-526ff9c0ed1f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 532.533339] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance b28550f5-4f8e-43ac-a020-c025586c6688 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 532.533339] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance ca158bd7-98e3-4ce0-9873-02571f3d9497 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 532.533484] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 532.533606] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 532.535248] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance e075eece-534b-4d97-8c82-abb960ef1abe actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 532.680801] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] Acquiring lock "759d9575-84bf-4dec-846a-e63126dae49c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.681158] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] Lock "759d9575-84bf-4dec-846a-e63126dae49c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 532.831423] env[62385]: DEBUG nova.network.neutron [-] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 533.039092] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 15703806-9f67-424d-aadb-634634810993 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 533.121828] env[62385]: DEBUG nova.network.neutron [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Successfully created port: 72a95deb-1344-4f59-9b77-a60b10ab47a8 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 533.335898] env[62385]: INFO nova.compute.manager [-] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Took 1.04 seconds to deallocate network for instance. [ 533.343839] env[62385]: DEBUG nova.compute.claims [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Aborting claim: {{(pid=62385) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 533.343929] env[62385]: DEBUG oslo_concurrency.lockutils [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.518129] env[62385]: DEBUG nova.compute.manager [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 533.543087] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 7ab1b11a-cf7c-4f6c-aeb2-c79681136721 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 533.557844] env[62385]: DEBUG nova.virt.hardware [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 533.558139] env[62385]: DEBUG nova.virt.hardware [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 533.558601] env[62385]: DEBUG nova.virt.hardware [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 533.558601] env[62385]: DEBUG nova.virt.hardware [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 533.558953] env[62385]: DEBUG nova.virt.hardware [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 533.558953] env[62385]: DEBUG nova.virt.hardware [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 533.558953] env[62385]: DEBUG nova.virt.hardware [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 533.559338] env[62385]: DEBUG nova.virt.hardware [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 533.559647] env[62385]: DEBUG nova.virt.hardware [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 533.559889] env[62385]: DEBUG nova.virt.hardware [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 533.560075] env[62385]: DEBUG nova.virt.hardware [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 533.561162] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a110c66c-d10a-4ddb-b0fc-539102000c88 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.574816] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6070182f-117d-4fd8-9ee2-9b784e4026e0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.923295] env[62385]: ERROR nova.compute.manager [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ed7ea7f6-f96a-40c0-86f0-88674a2954a4, please check neutron logs for more information. [ 533.923295] env[62385]: ERROR nova.compute.manager Traceback (most recent call last): [ 533.923295] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 533.923295] env[62385]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 533.923295] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 533.923295] env[62385]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 533.923295] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 533.923295] env[62385]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 533.923295] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.923295] env[62385]: ERROR nova.compute.manager self.force_reraise() [ 533.923295] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.923295] env[62385]: ERROR nova.compute.manager raise self.value [ 533.923295] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 533.923295] env[62385]: ERROR nova.compute.manager updated_port = self._update_port( [ 533.923295] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.923295] env[62385]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 533.924140] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.924140] env[62385]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 533.924140] env[62385]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ed7ea7f6-f96a-40c0-86f0-88674a2954a4, please check neutron logs for more information. [ 533.924140] env[62385]: ERROR nova.compute.manager [ 533.924140] env[62385]: Traceback (most recent call last): [ 533.924140] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 533.924140] env[62385]: listener.cb(fileno) [ 533.924140] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 533.924140] env[62385]: result = function(*args, **kwargs) [ 533.924140] env[62385]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 533.924140] env[62385]: return func(*args, **kwargs) [ 533.924140] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 533.924140] env[62385]: raise e [ 533.924140] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 533.924140] env[62385]: nwinfo = self.network_api.allocate_for_instance( [ 533.924140] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 533.924140] env[62385]: created_port_ids = self._update_ports_for_instance( [ 533.924140] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 533.924140] env[62385]: with excutils.save_and_reraise_exception(): [ 533.924140] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.924140] env[62385]: self.force_reraise() [ 533.924140] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.924140] env[62385]: raise self.value [ 533.924140] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 533.924140] env[62385]: updated_port = self._update_port( [ 533.924140] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.924140] env[62385]: _ensure_no_port_binding_failure(port) [ 533.924140] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.924140] env[62385]: raise exception.PortBindingFailed(port_id=port['id']) [ 533.924936] env[62385]: nova.exception.PortBindingFailed: Binding failed for port ed7ea7f6-f96a-40c0-86f0-88674a2954a4, please check neutron logs for more information. [ 533.924936] env[62385]: Removing descriptor: 19 [ 533.924936] env[62385]: ERROR nova.compute.manager [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ed7ea7f6-f96a-40c0-86f0-88674a2954a4, please check neutron logs for more information. [ 533.924936] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Traceback (most recent call last): [ 533.924936] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 533.924936] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] yield resources [ 533.924936] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 533.924936] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] self.driver.spawn(context, instance, image_meta, [ 533.924936] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 533.924936] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 533.924936] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 533.924936] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] vm_ref = self.build_virtual_machine(instance, [ 533.925332] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 533.925332] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] vif_infos = vmwarevif.get_vif_info(self._session, [ 533.925332] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 533.925332] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] for vif in network_info: [ 533.925332] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 533.925332] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] return self._sync_wrapper(fn, *args, **kwargs) [ 533.925332] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 533.925332] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] self.wait() [ 533.925332] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 533.925332] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] self[:] = self._gt.wait() [ 533.925332] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 533.925332] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] return self._exit_event.wait() [ 533.925332] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 533.925670] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] result = hub.switch() [ 533.925670] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 533.925670] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] return self.greenlet.switch() [ 533.925670] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 533.925670] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] result = function(*args, **kwargs) [ 533.925670] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 533.925670] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] return func(*args, **kwargs) [ 533.925670] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 533.925670] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] raise e [ 533.925670] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 533.925670] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] nwinfo = self.network_api.allocate_for_instance( [ 533.925670] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 533.925670] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] created_port_ids = self._update_ports_for_instance( [ 533.926037] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 533.926037] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] with excutils.save_and_reraise_exception(): [ 533.926037] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.926037] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] self.force_reraise() [ 533.926037] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.926037] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] raise self.value [ 533.926037] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 533.926037] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] updated_port = self._update_port( [ 533.926037] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.926037] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] _ensure_no_port_binding_failure(port) [ 533.926037] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.926037] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] raise exception.PortBindingFailed(port_id=port['id']) [ 533.926383] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] nova.exception.PortBindingFailed: Binding failed for port ed7ea7f6-f96a-40c0-86f0-88674a2954a4, please check neutron logs for more information. [ 533.926383] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] [ 533.926383] env[62385]: INFO nova.compute.manager [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Terminating instance [ 533.927645] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] Acquiring lock "refresh_cache-b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 533.927912] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] Acquired lock "refresh_cache-b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 533.927912] env[62385]: DEBUG nova.network.neutron [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 534.046326] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance d0f3ba00-5eee-4ff6-8999-3a318c9f0121 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 534.459305] env[62385]: DEBUG nova.network.neutron [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 534.555907] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 874d90b0-5657-467c-a934-31db63982ded has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 534.569206] env[62385]: DEBUG nova.network.neutron [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.946205] env[62385]: DEBUG oslo_concurrency.lockutils [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] Acquiring lock "be595c9e-703e-4246-ae1c-2c77b240cb19" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 534.946537] env[62385]: DEBUG oslo_concurrency.lockutils [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] Lock "be595c9e-703e-4246-ae1c-2c77b240cb19" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 535.005606] env[62385]: DEBUG nova.compute.manager [req-ec38ca35-3230-4a67-b7a5-53e63722f0b4 req-4bcda1e5-57ef-4ad0-a32c-a33785e21268 service nova] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Received event network-changed-9c5c2b66-3082-48a8-bec4-cc04e521703a {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 535.005791] env[62385]: DEBUG nova.compute.manager [req-ec38ca35-3230-4a67-b7a5-53e63722f0b4 req-4bcda1e5-57ef-4ad0-a32c-a33785e21268 service nova] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Refreshing instance network info cache due to event network-changed-9c5c2b66-3082-48a8-bec4-cc04e521703a. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 535.006011] env[62385]: DEBUG oslo_concurrency.lockutils [req-ec38ca35-3230-4a67-b7a5-53e63722f0b4 req-4bcda1e5-57ef-4ad0-a32c-a33785e21268 service nova] Acquiring lock "refresh_cache-b28550f5-4f8e-43ac-a020-c025586c6688" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 535.006166] env[62385]: DEBUG oslo_concurrency.lockutils [req-ec38ca35-3230-4a67-b7a5-53e63722f0b4 req-4bcda1e5-57ef-4ad0-a32c-a33785e21268 service nova] Acquired lock "refresh_cache-b28550f5-4f8e-43ac-a020-c025586c6688" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 535.006332] env[62385]: DEBUG nova.network.neutron [req-ec38ca35-3230-4a67-b7a5-53e63722f0b4 req-4bcda1e5-57ef-4ad0-a32c-a33785e21268 service nova] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Refreshing network info cache for port 9c5c2b66-3082-48a8-bec4-cc04e521703a {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 535.059629] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 2abff399-f03d-4c62-ac0b-440ba54bec29 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 535.072289] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] Releasing lock "refresh_cache-b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 535.072635] env[62385]: DEBUG nova.compute.manager [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 535.072824] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 535.073749] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a2fd2348-e5ad-4b3e-8632-e457e1e562d7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.087992] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eb26021-42fc-4d08-af11-67cdafaf161d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.111997] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a could not be found. [ 535.112557] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 535.112706] env[62385]: INFO nova.compute.manager [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 535.113026] env[62385]: DEBUG oslo.service.loopingcall [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 535.113177] env[62385]: DEBUG nova.compute.manager [-] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 535.113211] env[62385]: DEBUG nova.network.neutron [-] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 535.131466] env[62385]: DEBUG nova.network.neutron [-] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 535.190143] env[62385]: ERROR nova.compute.manager [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 72a95deb-1344-4f59-9b77-a60b10ab47a8, please check neutron logs for more information. [ 535.190143] env[62385]: ERROR nova.compute.manager Traceback (most recent call last): [ 535.190143] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 535.190143] env[62385]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 535.190143] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 535.190143] env[62385]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 535.190143] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 535.190143] env[62385]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 535.190143] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 535.190143] env[62385]: ERROR nova.compute.manager self.force_reraise() [ 535.190143] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 535.190143] env[62385]: ERROR nova.compute.manager raise self.value [ 535.190143] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 535.190143] env[62385]: ERROR nova.compute.manager updated_port = self._update_port( [ 535.190143] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 535.190143] env[62385]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 535.190604] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 535.190604] env[62385]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 535.190604] env[62385]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 72a95deb-1344-4f59-9b77-a60b10ab47a8, please check neutron logs for more information. [ 535.190604] env[62385]: ERROR nova.compute.manager [ 535.190604] env[62385]: Traceback (most recent call last): [ 535.190604] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 535.190604] env[62385]: listener.cb(fileno) [ 535.190604] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 535.190604] env[62385]: result = function(*args, **kwargs) [ 535.190604] env[62385]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 535.190604] env[62385]: return func(*args, **kwargs) [ 535.190604] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 535.190604] env[62385]: raise e [ 535.190604] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 535.190604] env[62385]: nwinfo = self.network_api.allocate_for_instance( [ 535.190604] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 535.190604] env[62385]: created_port_ids = self._update_ports_for_instance( [ 535.190604] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 535.190604] env[62385]: with excutils.save_and_reraise_exception(): [ 535.190604] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 535.190604] env[62385]: self.force_reraise() [ 535.190604] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 535.190604] env[62385]: raise self.value [ 535.190604] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 535.190604] env[62385]: updated_port = self._update_port( [ 535.190604] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 535.190604] env[62385]: _ensure_no_port_binding_failure(port) [ 535.190604] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 535.190604] env[62385]: raise exception.PortBindingFailed(port_id=port['id']) [ 535.191384] env[62385]: nova.exception.PortBindingFailed: Binding failed for port 72a95deb-1344-4f59-9b77-a60b10ab47a8, please check neutron logs for more information. [ 535.191384] env[62385]: Removing descriptor: 20 [ 535.191384] env[62385]: ERROR nova.compute.manager [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 72a95deb-1344-4f59-9b77-a60b10ab47a8, please check neutron logs for more information. [ 535.191384] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Traceback (most recent call last): [ 535.191384] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 535.191384] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] yield resources [ 535.191384] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 535.191384] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] self.driver.spawn(context, instance, image_meta, [ 535.191384] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 535.191384] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] self._vmops.spawn(context, instance, image_meta, injected_files, [ 535.191384] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 535.191384] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] vm_ref = self.build_virtual_machine(instance, [ 535.191747] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 535.191747] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] vif_infos = vmwarevif.get_vif_info(self._session, [ 535.191747] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 535.191747] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] for vif in network_info: [ 535.191747] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 535.191747] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] return self._sync_wrapper(fn, *args, **kwargs) [ 535.191747] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 535.191747] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] self.wait() [ 535.191747] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 535.191747] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] self[:] = self._gt.wait() [ 535.191747] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 535.191747] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] return self._exit_event.wait() [ 535.191747] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 535.192098] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] result = hub.switch() [ 535.192098] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 535.192098] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] return self.greenlet.switch() [ 535.192098] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 535.192098] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] result = function(*args, **kwargs) [ 535.192098] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 535.192098] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] return func(*args, **kwargs) [ 535.192098] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 535.192098] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] raise e [ 535.192098] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 535.192098] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] nwinfo = self.network_api.allocate_for_instance( [ 535.192098] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 535.192098] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] created_port_ids = self._update_ports_for_instance( [ 535.192491] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 535.192491] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] with excutils.save_and_reraise_exception(): [ 535.192491] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 535.192491] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] self.force_reraise() [ 535.192491] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 535.192491] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] raise self.value [ 535.192491] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 535.192491] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] updated_port = self._update_port( [ 535.192491] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 535.192491] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] _ensure_no_port_binding_failure(port) [ 535.192491] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 535.192491] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] raise exception.PortBindingFailed(port_id=port['id']) [ 535.192825] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] nova.exception.PortBindingFailed: Binding failed for port 72a95deb-1344-4f59-9b77-a60b10ab47a8, please check neutron logs for more information. [ 535.192825] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] [ 535.192825] env[62385]: INFO nova.compute.manager [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Terminating instance [ 535.193561] env[62385]: DEBUG oslo_concurrency.lockutils [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] Acquiring lock "refresh_cache-e075eece-534b-4d97-8c82-abb960ef1abe" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 535.193717] env[62385]: DEBUG oslo_concurrency.lockutils [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] Acquired lock "refresh_cache-e075eece-534b-4d97-8c82-abb960ef1abe" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 535.193980] env[62385]: DEBUG nova.network.neutron [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 535.317008] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Acquiring lock "e514a72d-6b4d-46f9-935b-3535bed4afc2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.317122] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Lock "e514a72d-6b4d-46f9-935b-3535bed4afc2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 535.533864] env[62385]: DEBUG nova.network.neutron [req-ec38ca35-3230-4a67-b7a5-53e63722f0b4 req-4bcda1e5-57ef-4ad0-a32c-a33785e21268 service nova] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 535.564681] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 9ddab157-12cf-4225-bc6f-e03ef9455d79 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 535.601216] env[62385]: DEBUG nova.network.neutron [req-ec38ca35-3230-4a67-b7a5-53e63722f0b4 req-4bcda1e5-57ef-4ad0-a32c-a33785e21268 service nova] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.633797] env[62385]: DEBUG nova.network.neutron [-] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.723253] env[62385]: DEBUG nova.network.neutron [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 535.813547] env[62385]: DEBUG nova.network.neutron [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 536.069844] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 759d9575-84bf-4dec-846a-e63126dae49c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 536.069844] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=62385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 536.069844] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1856MB phys_disk=200GB used_disk=7GB total_vcpus=48 used_vcpus=7 pci_stats=[] {{(pid=62385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 536.111062] env[62385]: DEBUG oslo_concurrency.lockutils [req-ec38ca35-3230-4a67-b7a5-53e63722f0b4 req-4bcda1e5-57ef-4ad0-a32c-a33785e21268 service nova] Releasing lock "refresh_cache-b28550f5-4f8e-43ac-a020-c025586c6688" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 536.111331] env[62385]: DEBUG nova.compute.manager [req-ec38ca35-3230-4a67-b7a5-53e63722f0b4 req-4bcda1e5-57ef-4ad0-a32c-a33785e21268 service nova] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Received event network-vif-deleted-9c5c2b66-3082-48a8-bec4-cc04e521703a {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 536.138089] env[62385]: INFO nova.compute.manager [-] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Took 1.02 seconds to deallocate network for instance. [ 536.144580] env[62385]: DEBUG nova.compute.claims [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Aborting claim: {{(pid=62385) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 536.144580] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 536.270788] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] Acquiring lock "0f1d83e7-bb40-4344-a344-2fc98efe8528" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 536.271031] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] Lock "0f1d83e7-bb40-4344-a344-2fc98efe8528" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 536.316572] env[62385]: DEBUG oslo_concurrency.lockutils [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] Releasing lock "refresh_cache-e075eece-534b-4d97-8c82-abb960ef1abe" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 536.317438] env[62385]: DEBUG nova.compute.manager [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 536.317438] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 536.317893] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1753e0dc-f8cb-469e-89d8-f04a1d729a2d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.330715] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73c15ce1-809e-4534-8177-cdda1df1081d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.356137] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e075eece-534b-4d97-8c82-abb960ef1abe could not be found. [ 536.356232] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 536.356441] env[62385]: INFO nova.compute.manager [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Took 0.04 seconds to destroy the instance on the hypervisor. [ 536.356689] env[62385]: DEBUG oslo.service.loopingcall [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 536.357961] env[62385]: DEBUG nova.compute.manager [-] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 536.358109] env[62385]: DEBUG nova.network.neutron [-] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 536.360527] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7f138a6-eecc-4218-96fa-fb814096df40 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.369153] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b87f741f-be69-44a9-ac30-ce1ef73295f3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.400890] env[62385]: DEBUG nova.network.neutron [-] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 536.402707] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afa32a5e-b3b2-4634-87f0-478e0f8d7634 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.411312] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ee74458-5b83-4e55-9013-a470115e9132 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.428277] env[62385]: DEBUG nova.compute.provider_tree [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 536.733470] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] Acquiring lock "0a21307b-74e7-4c5d-95a3-0800915c51ff" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 536.733707] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] Lock "0a21307b-74e7-4c5d-95a3-0800915c51ff" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 536.907271] env[62385]: DEBUG nova.network.neutron [-] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 536.933490] env[62385]: DEBUG nova.scheduler.client.report [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 537.411432] env[62385]: INFO nova.compute.manager [-] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Took 1.05 seconds to deallocate network for instance. [ 537.414708] env[62385]: DEBUG nova.compute.claims [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Aborting claim: {{(pid=62385) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 537.416931] env[62385]: DEBUG oslo_concurrency.lockutils [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.441020] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62385) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 537.441020] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 5.961s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 537.441020] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.037s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 537.441183] env[62385]: INFO nova.compute.claims [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] [instance: 15703806-9f67-424d-aadb-634634810993] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 537.568188] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "ce6b8f31-afee-4062-b860-d054c4d37be0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.568858] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "ce6b8f31-afee-4062-b860-d054c4d37be0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 538.594407] env[62385]: DEBUG nova.compute.manager [req-f8a61dd9-1c5e-4628-a4f6-6ddcd06f7537 req-a0b2c08e-b6c6-46aa-943e-6edd8ca9c957 service nova] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Received event network-changed-22d16c35-10f1-49bd-b538-c969c7034480 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 538.594407] env[62385]: DEBUG nova.compute.manager [req-f8a61dd9-1c5e-4628-a4f6-6ddcd06f7537 req-a0b2c08e-b6c6-46aa-943e-6edd8ca9c957 service nova] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Refreshing instance network info cache due to event network-changed-22d16c35-10f1-49bd-b538-c969c7034480. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 538.594407] env[62385]: DEBUG oslo_concurrency.lockutils [req-f8a61dd9-1c5e-4628-a4f6-6ddcd06f7537 req-a0b2c08e-b6c6-46aa-943e-6edd8ca9c957 service nova] Acquiring lock "refresh_cache-e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 538.594407] env[62385]: DEBUG oslo_concurrency.lockutils [req-f8a61dd9-1c5e-4628-a4f6-6ddcd06f7537 req-a0b2c08e-b6c6-46aa-943e-6edd8ca9c957 service nova] Acquired lock "refresh_cache-e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 538.596159] env[62385]: DEBUG nova.network.neutron [req-f8a61dd9-1c5e-4628-a4f6-6ddcd06f7537 req-a0b2c08e-b6c6-46aa-943e-6edd8ca9c957 service nova] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Refreshing network info cache for port 22d16c35-10f1-49bd-b538-c969c7034480 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 538.753370] env[62385]: DEBUG oslo_concurrency.lockutils [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Acquiring lock "4a99d2e8-7795-4016-ad6e-39a816c7fae1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.753618] env[62385]: DEBUG oslo_concurrency.lockutils [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Lock "4a99d2e8-7795-4016-ad6e-39a816c7fae1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 538.792896] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35851e2d-3646-4c7d-8ee0-229fe3cb4f75 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.805414] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f5eb967-0deb-417c-b8f6-f2f2ee52bf29 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.849429] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da877787-c12d-4bf7-b6b3-5638c9b9da47 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.857960] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b694d0e-a4df-4807-8af6-a23486c8ca62 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.876564] env[62385]: DEBUG nova.compute.provider_tree [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 538.892791] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] Acquiring lock "317e7e7c-7c6f-4855-a433-e55ea5571b92" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.893082] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] Lock "317e7e7c-7c6f-4855-a433-e55ea5571b92" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.124222] env[62385]: DEBUG nova.network.neutron [req-f8a61dd9-1c5e-4628-a4f6-6ddcd06f7537 req-a0b2c08e-b6c6-46aa-943e-6edd8ca9c957 service nova] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 539.317320] env[62385]: DEBUG nova.network.neutron [req-f8a61dd9-1c5e-4628-a4f6-6ddcd06f7537 req-a0b2c08e-b6c6-46aa-943e-6edd8ca9c957 service nova] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.383479] env[62385]: DEBUG nova.scheduler.client.report [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 539.821091] env[62385]: DEBUG oslo_concurrency.lockutils [req-f8a61dd9-1c5e-4628-a4f6-6ddcd06f7537 req-a0b2c08e-b6c6-46aa-943e-6edd8ca9c957 service nova] Releasing lock "refresh_cache-e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 539.821962] env[62385]: DEBUG nova.compute.manager [req-f8a61dd9-1c5e-4628-a4f6-6ddcd06f7537 req-a0b2c08e-b6c6-46aa-943e-6edd8ca9c957 service nova] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Received event network-vif-deleted-22d16c35-10f1-49bd-b538-c969c7034480 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 539.821962] env[62385]: DEBUG nova.compute.manager [req-f8a61dd9-1c5e-4628-a4f6-6ddcd06f7537 req-a0b2c08e-b6c6-46aa-943e-6edd8ca9c957 service nova] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Received event network-changed-ed7ea7f6-f96a-40c0-86f0-88674a2954a4 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 539.823048] env[62385]: DEBUG nova.compute.manager [req-f8a61dd9-1c5e-4628-a4f6-6ddcd06f7537 req-a0b2c08e-b6c6-46aa-943e-6edd8ca9c957 service nova] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Refreshing instance network info cache due to event network-changed-ed7ea7f6-f96a-40c0-86f0-88674a2954a4. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 539.823363] env[62385]: DEBUG oslo_concurrency.lockutils [req-f8a61dd9-1c5e-4628-a4f6-6ddcd06f7537 req-a0b2c08e-b6c6-46aa-943e-6edd8ca9c957 service nova] Acquiring lock "refresh_cache-b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 539.823534] env[62385]: DEBUG oslo_concurrency.lockutils [req-f8a61dd9-1c5e-4628-a4f6-6ddcd06f7537 req-a0b2c08e-b6c6-46aa-943e-6edd8ca9c957 service nova] Acquired lock "refresh_cache-b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 539.824913] env[62385]: DEBUG nova.network.neutron [req-f8a61dd9-1c5e-4628-a4f6-6ddcd06f7537 req-a0b2c08e-b6c6-46aa-943e-6edd8ca9c957 service nova] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Refreshing network info cache for port ed7ea7f6-f96a-40c0-86f0-88674a2954a4 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 539.888202] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.449s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 539.888743] env[62385]: DEBUG nova.compute.manager [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] [instance: 15703806-9f67-424d-aadb-634634810993] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 539.892378] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.885s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.363700] env[62385]: DEBUG nova.network.neutron [req-f8a61dd9-1c5e-4628-a4f6-6ddcd06f7537 req-a0b2c08e-b6c6-46aa-943e-6edd8ca9c957 service nova] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 540.396187] env[62385]: DEBUG nova.compute.utils [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 540.402411] env[62385]: DEBUG nova.compute.manager [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] [instance: 15703806-9f67-424d-aadb-634634810993] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 540.402592] env[62385]: DEBUG nova.network.neutron [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] [instance: 15703806-9f67-424d-aadb-634634810993] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 540.561188] env[62385]: DEBUG nova.network.neutron [req-f8a61dd9-1c5e-4628-a4f6-6ddcd06f7537 req-a0b2c08e-b6c6-46aa-943e-6edd8ca9c957 service nova] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.564828] env[62385]: DEBUG nova.policy [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8e4705092c3d470ca6f587f01d522e06', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3b5c6f7083c14a67967eaaa4b4ef5640', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 540.740508] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f504cb6-c0e9-4867-b42d-59e7e19ebf47 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.749051] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f168762b-4981-47ba-af3e-1d41413ec1da {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.784176] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90cba994-6df4-420c-a198-2ff74fd939a0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.795513] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2acbf29a-4e13-4129-9027-1c053d1684bb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.816137] env[62385]: DEBUG nova.compute.provider_tree [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 540.905585] env[62385]: DEBUG nova.compute.manager [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] [instance: 15703806-9f67-424d-aadb-634634810993] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 541.069098] env[62385]: DEBUG oslo_concurrency.lockutils [req-f8a61dd9-1c5e-4628-a4f6-6ddcd06f7537 req-a0b2c08e-b6c6-46aa-943e-6edd8ca9c957 service nova] Releasing lock "refresh_cache-b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 541.069098] env[62385]: DEBUG nova.compute.manager [req-f8a61dd9-1c5e-4628-a4f6-6ddcd06f7537 req-a0b2c08e-b6c6-46aa-943e-6edd8ca9c957 service nova] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Received event network-vif-deleted-ed7ea7f6-f96a-40c0-86f0-88674a2954a4 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 541.069098] env[62385]: DEBUG nova.compute.manager [req-f8a61dd9-1c5e-4628-a4f6-6ddcd06f7537 req-a0b2c08e-b6c6-46aa-943e-6edd8ca9c957 service nova] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Received event network-changed-72a95deb-1344-4f59-9b77-a60b10ab47a8 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 541.069098] env[62385]: DEBUG nova.compute.manager [req-f8a61dd9-1c5e-4628-a4f6-6ddcd06f7537 req-a0b2c08e-b6c6-46aa-943e-6edd8ca9c957 service nova] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Refreshing instance network info cache due to event network-changed-72a95deb-1344-4f59-9b77-a60b10ab47a8. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 541.069098] env[62385]: DEBUG oslo_concurrency.lockutils [req-f8a61dd9-1c5e-4628-a4f6-6ddcd06f7537 req-a0b2c08e-b6c6-46aa-943e-6edd8ca9c957 service nova] Acquiring lock "refresh_cache-e075eece-534b-4d97-8c82-abb960ef1abe" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 541.069379] env[62385]: DEBUG oslo_concurrency.lockutils [req-f8a61dd9-1c5e-4628-a4f6-6ddcd06f7537 req-a0b2c08e-b6c6-46aa-943e-6edd8ca9c957 service nova] Acquired lock "refresh_cache-e075eece-534b-4d97-8c82-abb960ef1abe" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 541.069379] env[62385]: DEBUG nova.network.neutron [req-f8a61dd9-1c5e-4628-a4f6-6ddcd06f7537 req-a0b2c08e-b6c6-46aa-943e-6edd8ca9c957 service nova] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Refreshing network info cache for port 72a95deb-1344-4f59-9b77-a60b10ab47a8 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 541.321057] env[62385]: DEBUG nova.scheduler.client.report [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 541.361262] env[62385]: DEBUG nova.network.neutron [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] [instance: 15703806-9f67-424d-aadb-634634810993] Successfully created port: 2ce7db43-289a-411f-8802-33921aed5c9e {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 541.614636] env[62385]: DEBUG nova.network.neutron [req-f8a61dd9-1c5e-4628-a4f6-6ddcd06f7537 req-a0b2c08e-b6c6-46aa-943e-6edd8ca9c957 service nova] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 541.826215] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.934s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 541.826916] env[62385]: ERROR nova.compute.manager [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 06b7091d-aaf7-4853-8e5c-1569a815efa0, please check neutron logs for more information. [ 541.826916] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Traceback (most recent call last): [ 541.826916] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 541.826916] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] self.driver.spawn(context, instance, image_meta, [ 541.826916] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 541.826916] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] self._vmops.spawn(context, instance, image_meta, injected_files, [ 541.826916] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 541.826916] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] vm_ref = self.build_virtual_machine(instance, [ 541.826916] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 541.826916] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] vif_infos = vmwarevif.get_vif_info(self._session, [ 541.826916] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 541.827364] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] for vif in network_info: [ 541.827364] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 541.827364] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] return self._sync_wrapper(fn, *args, **kwargs) [ 541.827364] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 541.827364] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] self.wait() [ 541.827364] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 541.827364] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] self[:] = self._gt.wait() [ 541.827364] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 541.827364] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] return self._exit_event.wait() [ 541.827364] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 541.827364] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] result = hub.switch() [ 541.827364] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 541.827364] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] return self.greenlet.switch() [ 541.827939] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 541.827939] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] result = function(*args, **kwargs) [ 541.827939] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 541.827939] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] return func(*args, **kwargs) [ 541.827939] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 541.827939] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] raise e [ 541.827939] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 541.827939] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] nwinfo = self.network_api.allocate_for_instance( [ 541.827939] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 541.827939] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] created_port_ids = self._update_ports_for_instance( [ 541.827939] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 541.827939] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] with excutils.save_and_reraise_exception(): [ 541.827939] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 541.828430] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] self.force_reraise() [ 541.828430] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 541.828430] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] raise self.value [ 541.828430] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 541.828430] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] updated_port = self._update_port( [ 541.828430] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 541.828430] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] _ensure_no_port_binding_failure(port) [ 541.828430] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 541.828430] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] raise exception.PortBindingFailed(port_id=port['id']) [ 541.828430] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] nova.exception.PortBindingFailed: Binding failed for port 06b7091d-aaf7-4853-8e5c-1569a815efa0, please check neutron logs for more information. [ 541.828430] env[62385]: ERROR nova.compute.manager [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] [ 541.829323] env[62385]: DEBUG nova.compute.utils [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Binding failed for port 06b7091d-aaf7-4853-8e5c-1569a815efa0, please check neutron logs for more information. {{(pid=62385) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 541.831282] env[62385]: DEBUG oslo_concurrency.lockutils [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.688s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.836100] env[62385]: DEBUG nova.compute.manager [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Build of instance d4f4eb24-ea78-4ab1-9786-818ea7769342 was re-scheduled: Binding failed for port 06b7091d-aaf7-4853-8e5c-1569a815efa0, please check neutron logs for more information. {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 541.836100] env[62385]: DEBUG nova.compute.manager [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Unplugging VIFs for instance {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 541.836100] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] Acquiring lock "refresh_cache-d4f4eb24-ea78-4ab1-9786-818ea7769342" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 541.836100] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] Acquired lock "refresh_cache-d4f4eb24-ea78-4ab1-9786-818ea7769342" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 541.836398] env[62385]: DEBUG nova.network.neutron [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 541.921162] env[62385]: DEBUG nova.compute.manager [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] [instance: 15703806-9f67-424d-aadb-634634810993] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 541.947681] env[62385]: DEBUG nova.network.neutron [req-f8a61dd9-1c5e-4628-a4f6-6ddcd06f7537 req-a0b2c08e-b6c6-46aa-943e-6edd8ca9c957 service nova] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 541.957811] env[62385]: DEBUG nova.virt.hardware [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 541.958677] env[62385]: DEBUG nova.virt.hardware [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 541.958677] env[62385]: DEBUG nova.virt.hardware [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 541.958677] env[62385]: DEBUG nova.virt.hardware [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 541.958677] env[62385]: DEBUG nova.virt.hardware [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 541.959378] env[62385]: DEBUG nova.virt.hardware [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 541.959378] env[62385]: DEBUG nova.virt.hardware [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 541.959378] env[62385]: DEBUG nova.virt.hardware [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 541.959595] env[62385]: DEBUG nova.virt.hardware [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 541.959690] env[62385]: DEBUG nova.virt.hardware [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 541.959952] env[62385]: DEBUG nova.virt.hardware [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 541.960911] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebb518fc-1e40-4c91-bb10-9dc3556f2f48 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.973331] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b22c4c94-b7c8-4436-8a3f-390d7369fd0f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.384491] env[62385]: DEBUG nova.network.neutron [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 542.450100] env[62385]: DEBUG oslo_concurrency.lockutils [req-f8a61dd9-1c5e-4628-a4f6-6ddcd06f7537 req-a0b2c08e-b6c6-46aa-943e-6edd8ca9c957 service nova] Releasing lock "refresh_cache-e075eece-534b-4d97-8c82-abb960ef1abe" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 542.455174] env[62385]: DEBUG nova.compute.manager [req-f8a61dd9-1c5e-4628-a4f6-6ddcd06f7537 req-a0b2c08e-b6c6-46aa-943e-6edd8ca9c957 service nova] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Received event network-vif-deleted-72a95deb-1344-4f59-9b77-a60b10ab47a8 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 542.588413] env[62385]: DEBUG nova.network.neutron [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 542.717038] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19ff04f6-ab11-43e4-975e-56ce6caee2b6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.726312] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29a8cc3d-4d91-4784-9bf0-85b2477ac0d6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.764930] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccdd7efb-5dfc-4e95-b422-6fafa532db26 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.773386] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fb0c03d-fccc-4ac6-a1b5-294cabd45c7e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.788353] env[62385]: DEBUG nova.compute.provider_tree [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 543.091085] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] Releasing lock "refresh_cache-d4f4eb24-ea78-4ab1-9786-818ea7769342" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 543.091652] env[62385]: DEBUG nova.compute.manager [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 543.091898] env[62385]: DEBUG nova.compute.manager [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 543.092095] env[62385]: DEBUG nova.network.neutron [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 543.135271] env[62385]: DEBUG nova.network.neutron [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 543.291196] env[62385]: DEBUG nova.scheduler.client.report [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 543.636074] env[62385]: DEBUG nova.network.neutron [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.694439] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Acquiring lock "0ba32627-c646-477d-9062-b84624d01e48" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.694439] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Lock "0ba32627-c646-477d-9062-b84624d01e48" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.803064] env[62385]: DEBUG oslo_concurrency.lockutils [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.968s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 543.803064] env[62385]: ERROR nova.compute.manager [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 85746a66-c8a0-4b2b-8298-31705b9ef890, please check neutron logs for more information. [ 543.803064] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Traceback (most recent call last): [ 543.803064] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 543.803064] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] self.driver.spawn(context, instance, image_meta, [ 543.803064] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 543.803064] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] self._vmops.spawn(context, instance, image_meta, injected_files, [ 543.803064] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 543.803064] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] vm_ref = self.build_virtual_machine(instance, [ 543.803421] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 543.803421] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] vif_infos = vmwarevif.get_vif_info(self._session, [ 543.803421] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 543.803421] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] for vif in network_info: [ 543.803421] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 543.803421] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] return self._sync_wrapper(fn, *args, **kwargs) [ 543.803421] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 543.803421] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] self.wait() [ 543.803421] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 543.803421] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] self[:] = self._gt.wait() [ 543.803421] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 543.803421] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] return self._exit_event.wait() [ 543.803421] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 543.803837] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] result = hub.switch() [ 543.803837] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 543.803837] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] return self.greenlet.switch() [ 543.803837] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 543.803837] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] result = function(*args, **kwargs) [ 543.803837] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 543.803837] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] return func(*args, **kwargs) [ 543.803837] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 543.803837] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] raise e [ 543.803837] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 543.803837] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] nwinfo = self.network_api.allocate_for_instance( [ 543.803837] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 543.803837] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] created_port_ids = self._update_ports_for_instance( [ 543.804290] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 543.804290] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] with excutils.save_and_reraise_exception(): [ 543.804290] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 543.804290] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] self.force_reraise() [ 543.804290] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 543.804290] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] raise self.value [ 543.804290] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 543.804290] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] updated_port = self._update_port( [ 543.804290] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 543.804290] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] _ensure_no_port_binding_failure(port) [ 543.804290] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 543.804290] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] raise exception.PortBindingFailed(port_id=port['id']) [ 543.804631] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] nova.exception.PortBindingFailed: Binding failed for port 85746a66-c8a0-4b2b-8298-31705b9ef890, please check neutron logs for more information. [ 543.804631] env[62385]: ERROR nova.compute.manager [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] [ 543.804631] env[62385]: DEBUG nova.compute.utils [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Binding failed for port 85746a66-c8a0-4b2b-8298-31705b9ef890, please check neutron logs for more information. {{(pid=62385) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 543.804631] env[62385]: DEBUG nova.compute.manager [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Build of instance ca158bd7-98e3-4ce0-9873-02571f3d9497 was re-scheduled: Binding failed for port 85746a66-c8a0-4b2b-8298-31705b9ef890, please check neutron logs for more information. {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 543.804631] env[62385]: DEBUG nova.compute.manager [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Unplugging VIFs for instance {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 543.805193] env[62385]: DEBUG oslo_concurrency.lockutils [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "refresh_cache-ca158bd7-98e3-4ce0-9873-02571f3d9497" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 543.805193] env[62385]: DEBUG oslo_concurrency.lockutils [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquired lock "refresh_cache-ca158bd7-98e3-4ce0-9873-02571f3d9497" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 543.805363] env[62385]: DEBUG nova.network.neutron [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 543.806518] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.860s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.139109] env[62385]: INFO nova.compute.manager [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] [instance: d4f4eb24-ea78-4ab1-9786-818ea7769342] Took 1.05 seconds to deallocate network for instance. [ 544.367399] env[62385]: DEBUG nova.compute.manager [req-a1253d29-aa6f-40d1-8777-50e39da08ab5 req-c79c9edb-a66f-4d6e-98fb-12ccc3118d93 service nova] [instance: 15703806-9f67-424d-aadb-634634810993] Received event network-changed-2ce7db43-289a-411f-8802-33921aed5c9e {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 544.367399] env[62385]: DEBUG nova.compute.manager [req-a1253d29-aa6f-40d1-8777-50e39da08ab5 req-c79c9edb-a66f-4d6e-98fb-12ccc3118d93 service nova] [instance: 15703806-9f67-424d-aadb-634634810993] Refreshing instance network info cache due to event network-changed-2ce7db43-289a-411f-8802-33921aed5c9e. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 544.367399] env[62385]: DEBUG oslo_concurrency.lockutils [req-a1253d29-aa6f-40d1-8777-50e39da08ab5 req-c79c9edb-a66f-4d6e-98fb-12ccc3118d93 service nova] Acquiring lock "refresh_cache-15703806-9f67-424d-aadb-634634810993" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 544.367399] env[62385]: DEBUG oslo_concurrency.lockutils [req-a1253d29-aa6f-40d1-8777-50e39da08ab5 req-c79c9edb-a66f-4d6e-98fb-12ccc3118d93 service nova] Acquired lock "refresh_cache-15703806-9f67-424d-aadb-634634810993" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 544.367399] env[62385]: DEBUG nova.network.neutron [req-a1253d29-aa6f-40d1-8777-50e39da08ab5 req-c79c9edb-a66f-4d6e-98fb-12ccc3118d93 service nova] [instance: 15703806-9f67-424d-aadb-634634810993] Refreshing network info cache for port 2ce7db43-289a-411f-8802-33921aed5c9e {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 544.373260] env[62385]: DEBUG nova.network.neutron [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 544.631688] env[62385]: ERROR nova.compute.manager [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2ce7db43-289a-411f-8802-33921aed5c9e, please check neutron logs for more information. [ 544.631688] env[62385]: ERROR nova.compute.manager Traceback (most recent call last): [ 544.631688] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 544.631688] env[62385]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 544.631688] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 544.631688] env[62385]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 544.631688] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 544.631688] env[62385]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 544.631688] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 544.631688] env[62385]: ERROR nova.compute.manager self.force_reraise() [ 544.631688] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 544.631688] env[62385]: ERROR nova.compute.manager raise self.value [ 544.631688] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 544.631688] env[62385]: ERROR nova.compute.manager updated_port = self._update_port( [ 544.631688] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 544.631688] env[62385]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 544.632422] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 544.632422] env[62385]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 544.632422] env[62385]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2ce7db43-289a-411f-8802-33921aed5c9e, please check neutron logs for more information. [ 544.632422] env[62385]: ERROR nova.compute.manager [ 544.632422] env[62385]: Traceback (most recent call last): [ 544.632422] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 544.632422] env[62385]: listener.cb(fileno) [ 544.632422] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 544.632422] env[62385]: result = function(*args, **kwargs) [ 544.632422] env[62385]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 544.632422] env[62385]: return func(*args, **kwargs) [ 544.632422] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 544.632422] env[62385]: raise e [ 544.632422] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 544.632422] env[62385]: nwinfo = self.network_api.allocate_for_instance( [ 544.632422] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 544.632422] env[62385]: created_port_ids = self._update_ports_for_instance( [ 544.632422] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 544.632422] env[62385]: with excutils.save_and_reraise_exception(): [ 544.632422] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 544.632422] env[62385]: self.force_reraise() [ 544.632422] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 544.632422] env[62385]: raise self.value [ 544.632422] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 544.632422] env[62385]: updated_port = self._update_port( [ 544.632422] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 544.632422] env[62385]: _ensure_no_port_binding_failure(port) [ 544.632422] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 544.632422] env[62385]: raise exception.PortBindingFailed(port_id=port['id']) [ 544.633203] env[62385]: nova.exception.PortBindingFailed: Binding failed for port 2ce7db43-289a-411f-8802-33921aed5c9e, please check neutron logs for more information. [ 544.633203] env[62385]: Removing descriptor: 19 [ 544.633203] env[62385]: ERROR nova.compute.manager [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] [instance: 15703806-9f67-424d-aadb-634634810993] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2ce7db43-289a-411f-8802-33921aed5c9e, please check neutron logs for more information. [ 544.633203] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] Traceback (most recent call last): [ 544.633203] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 544.633203] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] yield resources [ 544.633203] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 544.633203] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] self.driver.spawn(context, instance, image_meta, [ 544.633203] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 544.633203] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] self._vmops.spawn(context, instance, image_meta, injected_files, [ 544.633203] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 544.633203] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] vm_ref = self.build_virtual_machine(instance, [ 544.633606] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 544.633606] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] vif_infos = vmwarevif.get_vif_info(self._session, [ 544.633606] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 544.633606] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] for vif in network_info: [ 544.633606] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 544.633606] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] return self._sync_wrapper(fn, *args, **kwargs) [ 544.633606] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 544.633606] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] self.wait() [ 544.633606] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 544.633606] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] self[:] = self._gt.wait() [ 544.633606] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 544.633606] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] return self._exit_event.wait() [ 544.633606] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 544.633996] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] result = hub.switch() [ 544.633996] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 544.633996] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] return self.greenlet.switch() [ 544.633996] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 544.633996] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] result = function(*args, **kwargs) [ 544.633996] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 544.633996] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] return func(*args, **kwargs) [ 544.633996] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 544.633996] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] raise e [ 544.633996] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 544.633996] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] nwinfo = self.network_api.allocate_for_instance( [ 544.633996] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 544.633996] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] created_port_ids = self._update_ports_for_instance( [ 544.634412] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 544.634412] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] with excutils.save_and_reraise_exception(): [ 544.634412] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 544.634412] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] self.force_reraise() [ 544.634412] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 544.634412] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] raise self.value [ 544.634412] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 544.634412] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] updated_port = self._update_port( [ 544.634412] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 544.634412] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] _ensure_no_port_binding_failure(port) [ 544.634412] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 544.634412] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] raise exception.PortBindingFailed(port_id=port['id']) [ 544.635649] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] nova.exception.PortBindingFailed: Binding failed for port 2ce7db43-289a-411f-8802-33921aed5c9e, please check neutron logs for more information. [ 544.635649] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] [ 544.635649] env[62385]: INFO nova.compute.manager [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] [instance: 15703806-9f67-424d-aadb-634634810993] Terminating instance [ 544.635649] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] Acquiring lock "refresh_cache-15703806-9f67-424d-aadb-634634810993" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 544.666866] env[62385]: DEBUG nova.network.neutron [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 544.730029] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10a50668-50b0-40c6-89d0-a1ed866b6c19 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.745465] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0afaef5-6e1c-4343-8a02-6fa70de45516 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.782677] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2250488a-5462-4aef-8d1f-185302cde930 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.791540] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-074f20db-71d5-4198-be96-3fa9fb3a8897 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.807043] env[62385]: DEBUG nova.compute.provider_tree [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 544.919343] env[62385]: DEBUG nova.network.neutron [req-a1253d29-aa6f-40d1-8777-50e39da08ab5 req-c79c9edb-a66f-4d6e-98fb-12ccc3118d93 service nova] [instance: 15703806-9f67-424d-aadb-634634810993] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 545.173340] env[62385]: DEBUG oslo_concurrency.lockutils [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Releasing lock "refresh_cache-ca158bd7-98e3-4ce0-9873-02571f3d9497" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 545.173689] env[62385]: DEBUG nova.compute.manager [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 545.173805] env[62385]: DEBUG nova.compute.manager [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 545.173991] env[62385]: DEBUG nova.network.neutron [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 545.193108] env[62385]: INFO nova.scheduler.client.report [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] Deleted allocations for instance d4f4eb24-ea78-4ab1-9786-818ea7769342 [ 545.202970] env[62385]: DEBUG nova.network.neutron [req-a1253d29-aa6f-40d1-8777-50e39da08ab5 req-c79c9edb-a66f-4d6e-98fb-12ccc3118d93 service nova] [instance: 15703806-9f67-424d-aadb-634634810993] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.214061] env[62385]: DEBUG nova.network.neutron [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 545.312659] env[62385]: DEBUG nova.scheduler.client.report [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 545.458894] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Acquiring lock "f0f0eed0-a76a-42bc-9cca-91b93d7ad64c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.459148] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Lock "f0f0eed0-a76a-42bc-9cca-91b93d7ad64c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.707861] env[62385]: DEBUG oslo_concurrency.lockutils [req-a1253d29-aa6f-40d1-8777-50e39da08ab5 req-c79c9edb-a66f-4d6e-98fb-12ccc3118d93 service nova] Releasing lock "refresh_cache-15703806-9f67-424d-aadb-634634810993" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 545.708419] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0aa84dc2-388b-4d28-b968-c3a9fbe2d428 tempest-ServerDiagnosticsNegativeTest-29239534 tempest-ServerDiagnosticsNegativeTest-29239534-project-member] Lock "d4f4eb24-ea78-4ab1-9786-818ea7769342" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.918s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 545.708633] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] Acquired lock "refresh_cache-15703806-9f67-424d-aadb-634634810993" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 545.708841] env[62385]: DEBUG nova.network.neutron [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] [instance: 15703806-9f67-424d-aadb-634634810993] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 545.720300] env[62385]: DEBUG nova.network.neutron [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.821382] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.015s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 545.822026] env[62385]: ERROR nova.compute.manager [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 68a8e29c-58dc-4dbb-9f80-7dc9bbca6034, please check neutron logs for more information. [ 545.822026] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Traceback (most recent call last): [ 545.822026] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 545.822026] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] self.driver.spawn(context, instance, image_meta, [ 545.822026] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 545.822026] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 545.822026] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 545.822026] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] vm_ref = self.build_virtual_machine(instance, [ 545.822026] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 545.822026] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] vif_infos = vmwarevif.get_vif_info(self._session, [ 545.822026] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 545.822421] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] for vif in network_info: [ 545.822421] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 545.822421] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] return self._sync_wrapper(fn, *args, **kwargs) [ 545.822421] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 545.822421] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] self.wait() [ 545.822421] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 545.822421] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] self[:] = self._gt.wait() [ 545.822421] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 545.822421] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] return self._exit_event.wait() [ 545.822421] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 545.822421] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] result = hub.switch() [ 545.822421] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 545.822421] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] return self.greenlet.switch() [ 545.822922] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 545.822922] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] result = function(*args, **kwargs) [ 545.822922] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 545.822922] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] return func(*args, **kwargs) [ 545.822922] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 545.822922] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] raise e [ 545.822922] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 545.822922] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] nwinfo = self.network_api.allocate_for_instance( [ 545.822922] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 545.822922] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] created_port_ids = self._update_ports_for_instance( [ 545.822922] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 545.822922] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] with excutils.save_and_reraise_exception(): [ 545.822922] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 545.823393] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] self.force_reraise() [ 545.823393] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 545.823393] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] raise self.value [ 545.823393] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 545.823393] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] updated_port = self._update_port( [ 545.823393] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 545.823393] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] _ensure_no_port_binding_failure(port) [ 545.823393] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 545.823393] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] raise exception.PortBindingFailed(port_id=port['id']) [ 545.823393] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] nova.exception.PortBindingFailed: Binding failed for port 68a8e29c-58dc-4dbb-9f80-7dc9bbca6034, please check neutron logs for more information. [ 545.823393] env[62385]: ERROR nova.compute.manager [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] [ 545.823730] env[62385]: DEBUG nova.compute.utils [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Binding failed for port 68a8e29c-58dc-4dbb-9f80-7dc9bbca6034, please check neutron logs for more information. {{(pid=62385) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 545.830024] env[62385]: DEBUG nova.compute.manager [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Build of instance 077b25b0-3586-4c56-81fa-526ff9c0ed1f was re-scheduled: Binding failed for port 68a8e29c-58dc-4dbb-9f80-7dc9bbca6034, please check neutron logs for more information. {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 545.830024] env[62385]: DEBUG nova.compute.manager [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Unplugging VIFs for instance {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 545.830024] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Acquiring lock "refresh_cache-077b25b0-3586-4c56-81fa-526ff9c0ed1f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 545.830024] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Acquired lock "refresh_cache-077b25b0-3586-4c56-81fa-526ff9c0ed1f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 545.830390] env[62385]: DEBUG nova.network.neutron [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 545.830390] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.088s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.835152] env[62385]: INFO nova.compute.claims [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 546.221376] env[62385]: DEBUG nova.compute.manager [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 874d90b0-5657-467c-a934-31db63982ded] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 546.225029] env[62385]: INFO nova.compute.manager [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: ca158bd7-98e3-4ce0-9873-02571f3d9497] Took 1.05 seconds to deallocate network for instance. [ 546.310978] env[62385]: DEBUG nova.network.neutron [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] [instance: 15703806-9f67-424d-aadb-634634810993] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 546.437807] env[62385]: DEBUG nova.network.neutron [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 546.638991] env[62385]: DEBUG nova.network.neutron [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.651593] env[62385]: DEBUG nova.network.neutron [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] [instance: 15703806-9f67-424d-aadb-634634810993] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.759987] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.036620] env[62385]: DEBUG nova.compute.manager [req-64175e24-e749-4e10-991b-e71a23d6e6b6 req-57869370-81eb-4d34-a38e-ae263443c057 service nova] [instance: 15703806-9f67-424d-aadb-634634810993] Received event network-vif-deleted-2ce7db43-289a-411f-8802-33921aed5c9e {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 547.151440] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Releasing lock "refresh_cache-077b25b0-3586-4c56-81fa-526ff9c0ed1f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 547.151611] env[62385]: DEBUG nova.compute.manager [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 547.151611] env[62385]: DEBUG nova.compute.manager [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 547.151802] env[62385]: DEBUG nova.network.neutron [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 547.161995] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] Releasing lock "refresh_cache-15703806-9f67-424d-aadb-634634810993" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 547.162486] env[62385]: DEBUG nova.compute.manager [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] [instance: 15703806-9f67-424d-aadb-634634810993] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 547.162734] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] [instance: 15703806-9f67-424d-aadb-634634810993] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 547.163626] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d3dca3b9-8bf2-4805-91a9-d2f8e90bf839 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.173723] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8d12506-9e0d-41aa-b002-67ee94244e4b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.217955] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] [instance: 15703806-9f67-424d-aadb-634634810993] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 15703806-9f67-424d-aadb-634634810993 could not be found. [ 547.218368] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] [instance: 15703806-9f67-424d-aadb-634634810993] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 547.218668] env[62385]: INFO nova.compute.manager [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] [instance: 15703806-9f67-424d-aadb-634634810993] Took 0.06 seconds to destroy the instance on the hypervisor. [ 547.219052] env[62385]: DEBUG oslo.service.loopingcall [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 547.222876] env[62385]: DEBUG nova.compute.manager [-] [instance: 15703806-9f67-424d-aadb-634634810993] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 547.223176] env[62385]: DEBUG nova.network.neutron [-] [instance: 15703806-9f67-424d-aadb-634634810993] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 547.226679] env[62385]: DEBUG nova.network.neutron [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 547.288573] env[62385]: DEBUG nova.network.neutron [-] [instance: 15703806-9f67-424d-aadb-634634810993] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 547.291186] env[62385]: INFO nova.scheduler.client.report [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Deleted allocations for instance ca158bd7-98e3-4ce0-9873-02571f3d9497 [ 547.343456] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c4a5b21-5796-493b-81a9-efac2a69c8a6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.357558] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2be621f1-1f2b-4cb0-9b07-20e2093b2640 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.396919] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1386cb1-2838-4f0c-9f15-f9ad11493aa2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.406053] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdf5b0c4-410a-4c45-b18b-0c6255fb2326 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.427878] env[62385]: DEBUG nova.compute.provider_tree [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 547.731704] env[62385]: DEBUG nova.network.neutron [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 547.803516] env[62385]: DEBUG nova.network.neutron [-] [instance: 15703806-9f67-424d-aadb-634634810993] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 547.807696] env[62385]: DEBUG oslo_concurrency.lockutils [None req-acb4756d-8119-4c31-af17-e86b057fff54 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "ca158bd7-98e3-4ce0-9873-02571f3d9497" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.366s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 547.932317] env[62385]: DEBUG nova.scheduler.client.report [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 548.239025] env[62385]: INFO nova.compute.manager [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 077b25b0-3586-4c56-81fa-526ff9c0ed1f] Took 1.08 seconds to deallocate network for instance. [ 548.308237] env[62385]: INFO nova.compute.manager [-] [instance: 15703806-9f67-424d-aadb-634634810993] Took 1.08 seconds to deallocate network for instance. [ 548.311341] env[62385]: DEBUG nova.compute.manager [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 548.315308] env[62385]: DEBUG nova.compute.claims [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] [instance: 15703806-9f67-424d-aadb-634634810993] Aborting claim: {{(pid=62385) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 548.315308] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.438867] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.608s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 548.439486] env[62385]: DEBUG nova.compute.manager [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 548.446164] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.125s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.843647] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.949630] env[62385]: DEBUG nova.compute.utils [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 548.955623] env[62385]: DEBUG nova.compute.manager [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 548.955623] env[62385]: DEBUG nova.network.neutron [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 549.160036] env[62385]: DEBUG nova.policy [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5069827c870c4d0fa08acc5a619a6892', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '66d888fe9de54ee1b7c2c5947816694a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 549.290802] env[62385]: INFO nova.scheduler.client.report [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Deleted allocations for instance 077b25b0-3586-4c56-81fa-526ff9c0ed1f [ 549.443567] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3470128-43a5-4e14-9b9d-d59df386e676 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.451669] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3407c2d-0a0c-4000-b5bb-a87546a6c761 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.458183] env[62385]: DEBUG nova.compute.manager [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 549.504497] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32ed25fa-f0dd-4afe-971d-ea3c231f9ae4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.513809] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d8e8155-658c-40cc-91ba-c31dc3bf0d23 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.533523] env[62385]: DEBUG nova.compute.provider_tree [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 549.800645] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4eed3872-af67-4b52-b667-9b6107fff56e tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Lock "077b25b0-3586-4c56-81fa-526ff9c0ed1f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.287s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 550.040217] env[62385]: DEBUG nova.scheduler.client.report [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 550.307850] env[62385]: DEBUG nova.compute.manager [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 550.471354] env[62385]: DEBUG nova.compute.manager [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 550.508072] env[62385]: DEBUG nova.virt.hardware [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 550.508072] env[62385]: DEBUG nova.virt.hardware [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 550.508072] env[62385]: DEBUG nova.virt.hardware [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 550.508398] env[62385]: DEBUG nova.virt.hardware [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 550.508398] env[62385]: DEBUG nova.virt.hardware [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 550.508398] env[62385]: DEBUG nova.virt.hardware [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 550.508398] env[62385]: DEBUG nova.virt.hardware [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 550.508711] env[62385]: DEBUG nova.virt.hardware [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 550.509192] env[62385]: DEBUG nova.virt.hardware [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 550.509507] env[62385]: DEBUG nova.virt.hardware [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 550.509818] env[62385]: DEBUG nova.virt.hardware [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 550.510999] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d03af074-6781-4275-a0ad-9cc6b5dc1871 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.521566] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54a20e80-6151-4144-a964-3a6b6186562f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.552575] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.110s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 550.553259] env[62385]: ERROR nova.compute.manager [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9c5c2b66-3082-48a8-bec4-cc04e521703a, please check neutron logs for more information. [ 550.553259] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Traceback (most recent call last): [ 550.553259] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 550.553259] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] self.driver.spawn(context, instance, image_meta, [ 550.553259] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 550.553259] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] self._vmops.spawn(context, instance, image_meta, injected_files, [ 550.553259] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 550.553259] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] vm_ref = self.build_virtual_machine(instance, [ 550.553259] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 550.553259] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] vif_infos = vmwarevif.get_vif_info(self._session, [ 550.553259] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 550.553643] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] for vif in network_info: [ 550.553643] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 550.553643] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] return self._sync_wrapper(fn, *args, **kwargs) [ 550.553643] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 550.553643] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] self.wait() [ 550.553643] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 550.553643] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] self[:] = self._gt.wait() [ 550.553643] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 550.553643] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] return self._exit_event.wait() [ 550.553643] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 550.553643] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] result = hub.switch() [ 550.553643] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 550.553643] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] return self.greenlet.switch() [ 550.554057] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 550.554057] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] result = function(*args, **kwargs) [ 550.554057] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 550.554057] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] return func(*args, **kwargs) [ 550.554057] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 550.554057] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] raise e [ 550.554057] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 550.554057] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] nwinfo = self.network_api.allocate_for_instance( [ 550.554057] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 550.554057] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] created_port_ids = self._update_ports_for_instance( [ 550.554057] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 550.554057] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] with excutils.save_and_reraise_exception(): [ 550.554057] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 550.554481] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] self.force_reraise() [ 550.554481] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 550.554481] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] raise self.value [ 550.554481] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 550.554481] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] updated_port = self._update_port( [ 550.554481] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 550.554481] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] _ensure_no_port_binding_failure(port) [ 550.554481] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 550.554481] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] raise exception.PortBindingFailed(port_id=port['id']) [ 550.554481] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] nova.exception.PortBindingFailed: Binding failed for port 9c5c2b66-3082-48a8-bec4-cc04e521703a, please check neutron logs for more information. [ 550.554481] env[62385]: ERROR nova.compute.manager [instance: b28550f5-4f8e-43ac-a020-c025586c6688] [ 550.556821] env[62385]: DEBUG nova.compute.utils [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Binding failed for port 9c5c2b66-3082-48a8-bec4-cc04e521703a, please check neutron logs for more information. {{(pid=62385) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 550.556821] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.725s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.556821] env[62385]: INFO nova.compute.claims [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 550.563130] env[62385]: DEBUG nova.compute.manager [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Build of instance b28550f5-4f8e-43ac-a020-c025586c6688 was re-scheduled: Binding failed for port 9c5c2b66-3082-48a8-bec4-cc04e521703a, please check neutron logs for more information. {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 550.563816] env[62385]: DEBUG nova.compute.manager [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Unplugging VIFs for instance {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 550.563902] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Acquiring lock "refresh_cache-b28550f5-4f8e-43ac-a020-c025586c6688" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 550.563993] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Acquired lock "refresh_cache-b28550f5-4f8e-43ac-a020-c025586c6688" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 550.564343] env[62385]: DEBUG nova.network.neutron [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 550.855094] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.869174] env[62385]: DEBUG nova.network.neutron [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Successfully created port: 9997cf6d-ec9f-49f1-8ee9-d3684fbed8e8 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 551.185671] env[62385]: DEBUG nova.network.neutron [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 551.557290] env[62385]: DEBUG nova.network.neutron [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 551.984017] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66dabb42-bdd7-4638-815d-4514a344a1f1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.993475] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-518a4355-539e-4cf1-b950-38d7b9a0dec7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.031259] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8cdb888-7c97-46b3-8f8c-ebc32b0f21e8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.040439] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5247c3f-6e88-46ae-8e10-cf5bd24d3ad3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.055678] env[62385]: DEBUG nova.compute.provider_tree [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 552.059655] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Releasing lock "refresh_cache-b28550f5-4f8e-43ac-a020-c025586c6688" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 552.060128] env[62385]: DEBUG nova.compute.manager [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 552.060557] env[62385]: DEBUG nova.compute.manager [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 552.060557] env[62385]: DEBUG nova.network.neutron [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 552.121395] env[62385]: DEBUG nova.network.neutron [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 552.559559] env[62385]: DEBUG nova.scheduler.client.report [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 552.623391] env[62385]: DEBUG nova.network.neutron [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 553.068418] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.513s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 553.068951] env[62385]: DEBUG nova.compute.manager [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 553.072191] env[62385]: DEBUG oslo_concurrency.lockutils [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.728s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.128223] env[62385]: INFO nova.compute.manager [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: b28550f5-4f8e-43ac-a020-c025586c6688] Took 1.07 seconds to deallocate network for instance. [ 553.577856] env[62385]: DEBUG nova.compute.utils [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 553.579274] env[62385]: DEBUG nova.compute.manager [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 553.583059] env[62385]: DEBUG nova.network.neutron [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 553.776658] env[62385]: DEBUG nova.policy [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c7b4a66c838c411e877b805eaaa44775', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e97fb9f88a584182a2e441d77e69b523', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 553.830447] env[62385]: DEBUG oslo_concurrency.lockutils [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "01d010dc-6b5b-470b-a387-f95a17600e19" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.830822] env[62385]: DEBUG oslo_concurrency.lockutils [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "01d010dc-6b5b-470b-a387-f95a17600e19" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.841877] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] Acquiring lock "cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.842194] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] Lock "cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.995461] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76bac3b2-6e07-4069-b55c-9776203711ea {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.003216] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d29cdabe-a704-4595-be7a-01d393c753ba {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.034737] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2253c87-b6e0-4cce-98d6-7599f2a81552 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.042540] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-925316da-1028-40fc-9002-0e38a9883c59 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.057340] env[62385]: DEBUG nova.compute.provider_tree [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 554.086178] env[62385]: DEBUG nova.compute.manager [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 554.169959] env[62385]: INFO nova.scheduler.client.report [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Deleted allocations for instance b28550f5-4f8e-43ac-a020-c025586c6688 [ 554.459208] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Acquiring lock "70abd86c-aa63-4ccb-b185-65bf4977dfdf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.460829] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Lock "70abd86c-aa63-4ccb-b185-65bf4977dfdf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 554.563261] env[62385]: DEBUG nova.scheduler.client.report [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 554.681884] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bdef2272-50a5-47d8-abca-4e7868a35ba9 tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Lock "b28550f5-4f8e-43ac-a020-c025586c6688" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.823s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 555.067374] env[62385]: DEBUG oslo_concurrency.lockutils [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.995s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 555.067975] env[62385]: ERROR nova.compute.manager [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 22d16c35-10f1-49bd-b538-c969c7034480, please check neutron logs for more information. [ 555.067975] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Traceback (most recent call last): [ 555.067975] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 555.067975] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] self.driver.spawn(context, instance, image_meta, [ 555.067975] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 555.067975] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] self._vmops.spawn(context, instance, image_meta, injected_files, [ 555.067975] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 555.067975] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] vm_ref = self.build_virtual_machine(instance, [ 555.067975] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 555.067975] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] vif_infos = vmwarevif.get_vif_info(self._session, [ 555.067975] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 555.068383] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] for vif in network_info: [ 555.068383] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 555.068383] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] return self._sync_wrapper(fn, *args, **kwargs) [ 555.068383] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 555.068383] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] self.wait() [ 555.068383] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 555.068383] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] self[:] = self._gt.wait() [ 555.068383] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 555.068383] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] return self._exit_event.wait() [ 555.068383] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 555.068383] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] result = hub.switch() [ 555.068383] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 555.068383] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] return self.greenlet.switch() [ 555.068783] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 555.068783] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] result = function(*args, **kwargs) [ 555.068783] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 555.068783] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] return func(*args, **kwargs) [ 555.068783] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 555.068783] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] raise e [ 555.068783] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 555.068783] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] nwinfo = self.network_api.allocate_for_instance( [ 555.068783] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 555.068783] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] created_port_ids = self._update_ports_for_instance( [ 555.068783] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 555.068783] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] with excutils.save_and_reraise_exception(): [ 555.068783] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 555.069280] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] self.force_reraise() [ 555.069280] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 555.069280] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] raise self.value [ 555.069280] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 555.069280] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] updated_port = self._update_port( [ 555.069280] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 555.069280] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] _ensure_no_port_binding_failure(port) [ 555.069280] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 555.069280] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] raise exception.PortBindingFailed(port_id=port['id']) [ 555.069280] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] nova.exception.PortBindingFailed: Binding failed for port 22d16c35-10f1-49bd-b538-c969c7034480, please check neutron logs for more information. [ 555.069280] env[62385]: ERROR nova.compute.manager [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] [ 555.072612] env[62385]: DEBUG nova.compute.utils [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Binding failed for port 22d16c35-10f1-49bd-b538-c969c7034480, please check neutron logs for more information. {{(pid=62385) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 555.073946] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.929s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.084622] env[62385]: DEBUG nova.compute.manager [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Build of instance e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca was re-scheduled: Binding failed for port 22d16c35-10f1-49bd-b538-c969c7034480, please check neutron logs for more information. {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 555.084894] env[62385]: DEBUG nova.compute.manager [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Unplugging VIFs for instance {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 555.085157] env[62385]: DEBUG oslo_concurrency.lockutils [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] Acquiring lock "refresh_cache-e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 555.085309] env[62385]: DEBUG oslo_concurrency.lockutils [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] Acquired lock "refresh_cache-e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 555.085463] env[62385]: DEBUG nova.network.neutron [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 555.098902] env[62385]: DEBUG nova.compute.manager [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 555.123893] env[62385]: DEBUG nova.network.neutron [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Successfully created port: b093b332-dcaa-435e-8b66-3b749ce492d1 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 555.139762] env[62385]: DEBUG nova.virt.hardware [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 555.140149] env[62385]: DEBUG nova.virt.hardware [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 555.140393] env[62385]: DEBUG nova.virt.hardware [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 555.140705] env[62385]: DEBUG nova.virt.hardware [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 555.142954] env[62385]: DEBUG nova.virt.hardware [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 555.142954] env[62385]: DEBUG nova.virt.hardware [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 555.142954] env[62385]: DEBUG nova.virt.hardware [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 555.142954] env[62385]: DEBUG nova.virt.hardware [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 555.142954] env[62385]: DEBUG nova.virt.hardware [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 555.143336] env[62385]: DEBUG nova.virt.hardware [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 555.143336] env[62385]: DEBUG nova.virt.hardware [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 555.143732] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6c5c2af-d110-467c-b917-dd0085de71c7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.159783] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-556f1f39-3598-4df3-a642-227a4c316166 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.188770] env[62385]: DEBUG nova.compute.manager [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 555.510290] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] Acquiring lock "c348ef22-abb0-48cf-98a6-3a4b217dd186" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.510507] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] Lock "c348ef22-abb0-48cf-98a6-3a4b217dd186" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.566679] env[62385]: DEBUG oslo_concurrency.lockutils [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] Acquiring lock "ad12e5b0-6854-4281-9155-e1660393829c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.566899] env[62385]: DEBUG oslo_concurrency.lockutils [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] Lock "ad12e5b0-6854-4281-9155-e1660393829c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.681605] env[62385]: DEBUG nova.network.neutron [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 555.712875] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.061295] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e036236-a8ca-4a9f-b901-0f8605b11691 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.072630] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69b7602d-cdb9-4961-a404-7566200b1c53 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.115091] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7f274eb-4e81-4649-ba05-64fc37a40094 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.122303] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e98fda3-cc32-49a0-9378-e6f53c6b82a8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.140859] env[62385]: DEBUG nova.compute.provider_tree [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 556.399667] env[62385]: DEBUG nova.network.neutron [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.645733] env[62385]: DEBUG nova.scheduler.client.report [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 556.760045] env[62385]: ERROR nova.compute.manager [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9997cf6d-ec9f-49f1-8ee9-d3684fbed8e8, please check neutron logs for more information. [ 556.760045] env[62385]: ERROR nova.compute.manager Traceback (most recent call last): [ 556.760045] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 556.760045] env[62385]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 556.760045] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 556.760045] env[62385]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 556.760045] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 556.760045] env[62385]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 556.760045] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.760045] env[62385]: ERROR nova.compute.manager self.force_reraise() [ 556.760045] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.760045] env[62385]: ERROR nova.compute.manager raise self.value [ 556.760045] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 556.760045] env[62385]: ERROR nova.compute.manager updated_port = self._update_port( [ 556.760045] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.760045] env[62385]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 556.760696] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.760696] env[62385]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 556.760696] env[62385]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9997cf6d-ec9f-49f1-8ee9-d3684fbed8e8, please check neutron logs for more information. [ 556.760696] env[62385]: ERROR nova.compute.manager [ 556.760696] env[62385]: Traceback (most recent call last): [ 556.760696] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 556.760696] env[62385]: listener.cb(fileno) [ 556.760696] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 556.760696] env[62385]: result = function(*args, **kwargs) [ 556.760696] env[62385]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 556.760696] env[62385]: return func(*args, **kwargs) [ 556.760696] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 556.760696] env[62385]: raise e [ 556.760696] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 556.760696] env[62385]: nwinfo = self.network_api.allocate_for_instance( [ 556.760696] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 556.760696] env[62385]: created_port_ids = self._update_ports_for_instance( [ 556.760696] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 556.760696] env[62385]: with excutils.save_and_reraise_exception(): [ 556.760696] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.760696] env[62385]: self.force_reraise() [ 556.760696] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.760696] env[62385]: raise self.value [ 556.760696] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 556.760696] env[62385]: updated_port = self._update_port( [ 556.760696] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.760696] env[62385]: _ensure_no_port_binding_failure(port) [ 556.760696] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.760696] env[62385]: raise exception.PortBindingFailed(port_id=port['id']) [ 556.761491] env[62385]: nova.exception.PortBindingFailed: Binding failed for port 9997cf6d-ec9f-49f1-8ee9-d3684fbed8e8, please check neutron logs for more information. [ 556.761491] env[62385]: Removing descriptor: 20 [ 556.761491] env[62385]: ERROR nova.compute.manager [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9997cf6d-ec9f-49f1-8ee9-d3684fbed8e8, please check neutron logs for more information. [ 556.761491] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Traceback (most recent call last): [ 556.761491] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 556.761491] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] yield resources [ 556.761491] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 556.761491] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] self.driver.spawn(context, instance, image_meta, [ 556.761491] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 556.761491] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] self._vmops.spawn(context, instance, image_meta, injected_files, [ 556.761491] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 556.761491] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] vm_ref = self.build_virtual_machine(instance, [ 556.761862] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 556.761862] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] vif_infos = vmwarevif.get_vif_info(self._session, [ 556.761862] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 556.761862] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] for vif in network_info: [ 556.761862] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 556.761862] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] return self._sync_wrapper(fn, *args, **kwargs) [ 556.761862] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 556.761862] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] self.wait() [ 556.761862] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 556.761862] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] self[:] = self._gt.wait() [ 556.761862] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 556.761862] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] return self._exit_event.wait() [ 556.761862] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 556.762241] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] result = hub.switch() [ 556.762241] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 556.762241] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] return self.greenlet.switch() [ 556.762241] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 556.762241] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] result = function(*args, **kwargs) [ 556.762241] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 556.762241] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] return func(*args, **kwargs) [ 556.762241] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 556.762241] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] raise e [ 556.762241] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 556.762241] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] nwinfo = self.network_api.allocate_for_instance( [ 556.762241] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 556.762241] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] created_port_ids = self._update_ports_for_instance( [ 556.762667] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 556.762667] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] with excutils.save_and_reraise_exception(): [ 556.762667] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.762667] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] self.force_reraise() [ 556.762667] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.762667] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] raise self.value [ 556.762667] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 556.762667] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] updated_port = self._update_port( [ 556.762667] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.762667] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] _ensure_no_port_binding_failure(port) [ 556.762667] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.762667] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] raise exception.PortBindingFailed(port_id=port['id']) [ 556.766140] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] nova.exception.PortBindingFailed: Binding failed for port 9997cf6d-ec9f-49f1-8ee9-d3684fbed8e8, please check neutron logs for more information. [ 556.766140] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] [ 556.766140] env[62385]: INFO nova.compute.manager [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Terminating instance [ 556.766140] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Acquiring lock "refresh_cache-7ab1b11a-cf7c-4f6c-aeb2-c79681136721" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 556.766140] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Acquired lock "refresh_cache-7ab1b11a-cf7c-4f6c-aeb2-c79681136721" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 556.766140] env[62385]: DEBUG nova.network.neutron [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 556.887888] env[62385]: DEBUG nova.compute.manager [req-20f6e950-b415-4872-ba27-3dcb9d8931fb req-63902bfe-119e-4f83-98df-8a53bf34040c service nova] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Received event network-changed-9997cf6d-ec9f-49f1-8ee9-d3684fbed8e8 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 556.888269] env[62385]: DEBUG nova.compute.manager [req-20f6e950-b415-4872-ba27-3dcb9d8931fb req-63902bfe-119e-4f83-98df-8a53bf34040c service nova] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Refreshing instance network info cache due to event network-changed-9997cf6d-ec9f-49f1-8ee9-d3684fbed8e8. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 556.888563] env[62385]: DEBUG oslo_concurrency.lockutils [req-20f6e950-b415-4872-ba27-3dcb9d8931fb req-63902bfe-119e-4f83-98df-8a53bf34040c service nova] Acquiring lock "refresh_cache-7ab1b11a-cf7c-4f6c-aeb2-c79681136721" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 556.906022] env[62385]: DEBUG oslo_concurrency.lockutils [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] Releasing lock "refresh_cache-e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 556.906022] env[62385]: DEBUG nova.compute.manager [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 556.906022] env[62385]: DEBUG nova.compute.manager [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 556.906022] env[62385]: DEBUG nova.network.neutron [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 556.962131] env[62385]: DEBUG nova.network.neutron [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 557.155358] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.082s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 557.155895] env[62385]: ERROR nova.compute.manager [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ed7ea7f6-f96a-40c0-86f0-88674a2954a4, please check neutron logs for more information. [ 557.155895] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Traceback (most recent call last): [ 557.155895] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 557.155895] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] self.driver.spawn(context, instance, image_meta, [ 557.155895] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 557.155895] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 557.155895] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 557.155895] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] vm_ref = self.build_virtual_machine(instance, [ 557.155895] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 557.155895] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] vif_infos = vmwarevif.get_vif_info(self._session, [ 557.155895] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 557.156278] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] for vif in network_info: [ 557.156278] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 557.156278] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] return self._sync_wrapper(fn, *args, **kwargs) [ 557.156278] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 557.156278] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] self.wait() [ 557.156278] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 557.156278] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] self[:] = self._gt.wait() [ 557.156278] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 557.156278] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] return self._exit_event.wait() [ 557.156278] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 557.156278] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] result = hub.switch() [ 557.156278] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 557.156278] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] return self.greenlet.switch() [ 557.156635] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 557.156635] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] result = function(*args, **kwargs) [ 557.156635] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 557.156635] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] return func(*args, **kwargs) [ 557.156635] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 557.156635] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] raise e [ 557.156635] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 557.156635] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] nwinfo = self.network_api.allocate_for_instance( [ 557.156635] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 557.156635] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] created_port_ids = self._update_ports_for_instance( [ 557.156635] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 557.156635] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] with excutils.save_and_reraise_exception(): [ 557.156635] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.156982] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] self.force_reraise() [ 557.156982] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.156982] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] raise self.value [ 557.156982] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 557.156982] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] updated_port = self._update_port( [ 557.156982] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.156982] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] _ensure_no_port_binding_failure(port) [ 557.156982] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.156982] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] raise exception.PortBindingFailed(port_id=port['id']) [ 557.156982] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] nova.exception.PortBindingFailed: Binding failed for port ed7ea7f6-f96a-40c0-86f0-88674a2954a4, please check neutron logs for more information. [ 557.156982] env[62385]: ERROR nova.compute.manager [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] [ 557.158834] env[62385]: DEBUG nova.compute.utils [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Binding failed for port ed7ea7f6-f96a-40c0-86f0-88674a2954a4, please check neutron logs for more information. {{(pid=62385) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 557.160802] env[62385]: DEBUG oslo_concurrency.lockutils [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.746s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.164318] env[62385]: DEBUG nova.compute.manager [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Build of instance b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a was re-scheduled: Binding failed for port ed7ea7f6-f96a-40c0-86f0-88674a2954a4, please check neutron logs for more information. {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 557.165186] env[62385]: DEBUG nova.compute.manager [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Unplugging VIFs for instance {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 557.165623] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] Acquiring lock "refresh_cache-b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 557.165817] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] Acquired lock "refresh_cache-b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 557.166051] env[62385]: DEBUG nova.network.neutron [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 557.314970] env[62385]: DEBUG nova.network.neutron [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 557.464240] env[62385]: DEBUG nova.network.neutron [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.749222] env[62385]: DEBUG nova.network.neutron [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 557.812408] env[62385]: DEBUG nova.network.neutron [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.970932] env[62385]: INFO nova.compute.manager [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] [instance: e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca] Took 1.07 seconds to deallocate network for instance. [ 558.130983] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dc2ffaf-5f69-4d5c-9335-bddc8e29727c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.141976] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89d48201-1529-4e16-af97-ce607238e4d4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.176017] env[62385]: DEBUG nova.network.neutron [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.177525] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a0523d3-5b43-4082-9435-8de8a4c21456 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.185528] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ff6fd0a-b544-43c7-a2a1-e240c4d3196b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.208031] env[62385]: DEBUG nova.compute.provider_tree [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 558.319186] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Releasing lock "refresh_cache-7ab1b11a-cf7c-4f6c-aeb2-c79681136721" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 558.323012] env[62385]: DEBUG nova.compute.manager [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 558.323246] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 558.323578] env[62385]: DEBUG oslo_concurrency.lockutils [req-20f6e950-b415-4872-ba27-3dcb9d8931fb req-63902bfe-119e-4f83-98df-8a53bf34040c service nova] Acquired lock "refresh_cache-7ab1b11a-cf7c-4f6c-aeb2-c79681136721" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 558.323756] env[62385]: DEBUG nova.network.neutron [req-20f6e950-b415-4872-ba27-3dcb9d8931fb req-63902bfe-119e-4f83-98df-8a53bf34040c service nova] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Refreshing network info cache for port 9997cf6d-ec9f-49f1-8ee9-d3684fbed8e8 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 558.325719] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-70cb5a8b-3b63-4f56-9401-737761450f6f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.337227] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9fcf6e5-44b8-44e7-94d6-1a0206dc11aa {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.362665] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7ab1b11a-cf7c-4f6c-aeb2-c79681136721 could not be found. [ 558.362881] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 558.363067] env[62385]: INFO nova.compute.manager [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Took 0.04 seconds to destroy the instance on the hypervisor. [ 558.363311] env[62385]: DEBUG oslo.service.loopingcall [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 558.363522] env[62385]: DEBUG nova.compute.manager [-] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 558.363618] env[62385]: DEBUG nova.network.neutron [-] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 558.418395] env[62385]: DEBUG nova.network.neutron [-] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 558.682626] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] Releasing lock "refresh_cache-b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 558.682865] env[62385]: DEBUG nova.compute.manager [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 558.683037] env[62385]: DEBUG nova.compute.manager [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 558.683211] env[62385]: DEBUG nova.network.neutron [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 558.713175] env[62385]: DEBUG nova.scheduler.client.report [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 558.722207] env[62385]: DEBUG nova.network.neutron [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 558.880053] env[62385]: DEBUG nova.network.neutron [req-20f6e950-b415-4872-ba27-3dcb9d8931fb req-63902bfe-119e-4f83-98df-8a53bf34040c service nova] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 558.920803] env[62385]: DEBUG nova.network.neutron [-] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 559.003452] env[62385]: INFO nova.scheduler.client.report [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] Deleted allocations for instance e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca [ 559.222496] env[62385]: DEBUG oslo_concurrency.lockutils [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.061s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 559.224595] env[62385]: ERROR nova.compute.manager [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 72a95deb-1344-4f59-9b77-a60b10ab47a8, please check neutron logs for more information. [ 559.224595] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Traceback (most recent call last): [ 559.224595] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 559.224595] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] self.driver.spawn(context, instance, image_meta, [ 559.224595] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 559.224595] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] self._vmops.spawn(context, instance, image_meta, injected_files, [ 559.224595] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 559.224595] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] vm_ref = self.build_virtual_machine(instance, [ 559.224595] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 559.224595] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] vif_infos = vmwarevif.get_vif_info(self._session, [ 559.224595] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 559.224974] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] for vif in network_info: [ 559.224974] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 559.224974] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] return self._sync_wrapper(fn, *args, **kwargs) [ 559.224974] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 559.224974] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] self.wait() [ 559.224974] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 559.224974] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] self[:] = self._gt.wait() [ 559.224974] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 559.224974] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] return self._exit_event.wait() [ 559.224974] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 559.224974] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] result = hub.switch() [ 559.224974] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 559.224974] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] return self.greenlet.switch() [ 559.226033] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 559.226033] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] result = function(*args, **kwargs) [ 559.226033] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 559.226033] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] return func(*args, **kwargs) [ 559.226033] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 559.226033] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] raise e [ 559.226033] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 559.226033] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] nwinfo = self.network_api.allocate_for_instance( [ 559.226033] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 559.226033] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] created_port_ids = self._update_ports_for_instance( [ 559.226033] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 559.226033] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] with excutils.save_and_reraise_exception(): [ 559.226033] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 559.226445] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] self.force_reraise() [ 559.226445] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 559.226445] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] raise self.value [ 559.226445] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 559.226445] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] updated_port = self._update_port( [ 559.226445] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 559.226445] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] _ensure_no_port_binding_failure(port) [ 559.226445] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 559.226445] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] raise exception.PortBindingFailed(port_id=port['id']) [ 559.226445] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] nova.exception.PortBindingFailed: Binding failed for port 72a95deb-1344-4f59-9b77-a60b10ab47a8, please check neutron logs for more information. [ 559.226445] env[62385]: ERROR nova.compute.manager [instance: e075eece-534b-4d97-8c82-abb960ef1abe] [ 559.226803] env[62385]: DEBUG nova.compute.utils [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Binding failed for port 72a95deb-1344-4f59-9b77-a60b10ab47a8, please check neutron logs for more information. {{(pid=62385) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 559.226803] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.467s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 559.231820] env[62385]: INFO nova.compute.claims [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 874d90b0-5657-467c-a934-31db63982ded] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 559.232372] env[62385]: DEBUG nova.network.neutron [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 559.233464] env[62385]: DEBUG nova.compute.manager [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Build of instance e075eece-534b-4d97-8c82-abb960ef1abe was re-scheduled: Binding failed for port 72a95deb-1344-4f59-9b77-a60b10ab47a8, please check neutron logs for more information. {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 559.235974] env[62385]: DEBUG nova.compute.manager [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Unplugging VIFs for instance {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 559.235974] env[62385]: DEBUG oslo_concurrency.lockutils [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] Acquiring lock "refresh_cache-e075eece-534b-4d97-8c82-abb960ef1abe" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 559.235974] env[62385]: DEBUG oslo_concurrency.lockutils [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] Acquired lock "refresh_cache-e075eece-534b-4d97-8c82-abb960ef1abe" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 559.235974] env[62385]: DEBUG nova.network.neutron [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 559.246130] env[62385]: DEBUG nova.network.neutron [req-20f6e950-b415-4872-ba27-3dcb9d8931fb req-63902bfe-119e-4f83-98df-8a53bf34040c service nova] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 559.424536] env[62385]: INFO nova.compute.manager [-] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Took 1.06 seconds to deallocate network for instance. [ 559.427857] env[62385]: DEBUG nova.compute.claims [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Aborting claim: {{(pid=62385) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 559.428217] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.510037] env[62385]: DEBUG nova.compute.manager [req-f79ec5ec-7907-4727-a2d0-4281339d5359 req-523b7606-e83b-4382-8c58-748e7d7c6412 service nova] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Received event network-vif-deleted-9997cf6d-ec9f-49f1-8ee9-d3684fbed8e8 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 559.516628] env[62385]: DEBUG oslo_concurrency.lockutils [None req-debb350d-79c4-4532-9df3-021f71520245 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737 tempest-FloatingIPsAssociationNegativeTestJSON-1694955737-project-member] Lock "e5fddcfc-df93-4c2b-ad5d-0d2f52b10aca" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.320s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 559.738871] env[62385]: INFO nova.compute.manager [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] [instance: b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a] Took 1.06 seconds to deallocate network for instance. [ 559.751881] env[62385]: DEBUG oslo_concurrency.lockutils [req-20f6e950-b415-4872-ba27-3dcb9d8931fb req-63902bfe-119e-4f83-98df-8a53bf34040c service nova] Releasing lock "refresh_cache-7ab1b11a-cf7c-4f6c-aeb2-c79681136721" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 559.789996] env[62385]: DEBUG nova.network.neutron [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 560.019979] env[62385]: DEBUG nova.compute.manager [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 560.029390] env[62385]: DEBUG nova.network.neutron [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.316345] env[62385]: ERROR nova.compute.manager [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b093b332-dcaa-435e-8b66-3b749ce492d1, please check neutron logs for more information. [ 560.316345] env[62385]: ERROR nova.compute.manager Traceback (most recent call last): [ 560.316345] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 560.316345] env[62385]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 560.316345] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 560.316345] env[62385]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 560.316345] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 560.316345] env[62385]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 560.316345] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 560.316345] env[62385]: ERROR nova.compute.manager self.force_reraise() [ 560.316345] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 560.316345] env[62385]: ERROR nova.compute.manager raise self.value [ 560.316345] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 560.316345] env[62385]: ERROR nova.compute.manager updated_port = self._update_port( [ 560.316345] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 560.316345] env[62385]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 560.316872] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 560.316872] env[62385]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 560.316872] env[62385]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b093b332-dcaa-435e-8b66-3b749ce492d1, please check neutron logs for more information. [ 560.316872] env[62385]: ERROR nova.compute.manager [ 560.316872] env[62385]: Traceback (most recent call last): [ 560.316872] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 560.316872] env[62385]: listener.cb(fileno) [ 560.316872] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 560.316872] env[62385]: result = function(*args, **kwargs) [ 560.316872] env[62385]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 560.316872] env[62385]: return func(*args, **kwargs) [ 560.316872] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 560.316872] env[62385]: raise e [ 560.316872] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 560.316872] env[62385]: nwinfo = self.network_api.allocate_for_instance( [ 560.316872] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 560.316872] env[62385]: created_port_ids = self._update_ports_for_instance( [ 560.316872] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 560.316872] env[62385]: with excutils.save_and_reraise_exception(): [ 560.316872] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 560.316872] env[62385]: self.force_reraise() [ 560.316872] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 560.316872] env[62385]: raise self.value [ 560.316872] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 560.316872] env[62385]: updated_port = self._update_port( [ 560.316872] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 560.316872] env[62385]: _ensure_no_port_binding_failure(port) [ 560.316872] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 560.316872] env[62385]: raise exception.PortBindingFailed(port_id=port['id']) [ 560.317651] env[62385]: nova.exception.PortBindingFailed: Binding failed for port b093b332-dcaa-435e-8b66-3b749ce492d1, please check neutron logs for more information. [ 560.317651] env[62385]: Removing descriptor: 19 [ 560.317651] env[62385]: ERROR nova.compute.manager [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b093b332-dcaa-435e-8b66-3b749ce492d1, please check neutron logs for more information. [ 560.317651] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Traceback (most recent call last): [ 560.317651] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 560.317651] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] yield resources [ 560.317651] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 560.317651] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] self.driver.spawn(context, instance, image_meta, [ 560.317651] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 560.317651] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] self._vmops.spawn(context, instance, image_meta, injected_files, [ 560.317651] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 560.317651] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] vm_ref = self.build_virtual_machine(instance, [ 560.318031] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 560.318031] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] vif_infos = vmwarevif.get_vif_info(self._session, [ 560.318031] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 560.318031] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] for vif in network_info: [ 560.318031] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 560.318031] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] return self._sync_wrapper(fn, *args, **kwargs) [ 560.318031] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 560.318031] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] self.wait() [ 560.318031] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 560.318031] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] self[:] = self._gt.wait() [ 560.318031] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 560.318031] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] return self._exit_event.wait() [ 560.318031] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 560.318406] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] result = hub.switch() [ 560.318406] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 560.318406] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] return self.greenlet.switch() [ 560.318406] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 560.318406] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] result = function(*args, **kwargs) [ 560.318406] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 560.318406] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] return func(*args, **kwargs) [ 560.318406] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 560.318406] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] raise e [ 560.318406] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 560.318406] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] nwinfo = self.network_api.allocate_for_instance( [ 560.318406] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 560.318406] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] created_port_ids = self._update_ports_for_instance( [ 560.318807] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 560.318807] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] with excutils.save_and_reraise_exception(): [ 560.318807] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 560.318807] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] self.force_reraise() [ 560.318807] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 560.318807] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] raise self.value [ 560.318807] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 560.318807] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] updated_port = self._update_port( [ 560.318807] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 560.318807] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] _ensure_no_port_binding_failure(port) [ 560.318807] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 560.318807] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] raise exception.PortBindingFailed(port_id=port['id']) [ 560.319244] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] nova.exception.PortBindingFailed: Binding failed for port b093b332-dcaa-435e-8b66-3b749ce492d1, please check neutron logs for more information. [ 560.319244] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] [ 560.319244] env[62385]: INFO nova.compute.manager [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Terminating instance [ 560.323087] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] Acquiring lock "refresh_cache-d0f3ba00-5eee-4ff6-8999-3a318c9f0121" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 560.323087] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] Acquired lock "refresh_cache-d0f3ba00-5eee-4ff6-8999-3a318c9f0121" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 560.323087] env[62385]: DEBUG nova.network.neutron [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 560.536055] env[62385]: DEBUG oslo_concurrency.lockutils [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] Releasing lock "refresh_cache-e075eece-534b-4d97-8c82-abb960ef1abe" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 560.536782] env[62385]: DEBUG nova.compute.manager [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 560.536782] env[62385]: DEBUG nova.compute.manager [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 560.536782] env[62385]: DEBUG nova.network.neutron [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 560.547194] env[62385]: DEBUG oslo_concurrency.lockutils [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.589413] env[62385]: DEBUG nova.network.neutron [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 560.694610] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d785afa0-5db2-4047-963f-d0660937a428 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.702517] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c00b276-5ae5-4cd3-b30a-b7c38e9f83e3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.742575] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a09b38f-7704-4f2d-8ad1-fba0fc14a21d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.756927] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3292f619-1a6a-429a-9186-f310af9e34f3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.775353] env[62385]: DEBUG nova.compute.provider_tree [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 560.791073] env[62385]: INFO nova.scheduler.client.report [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] Deleted allocations for instance b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a [ 560.847943] env[62385]: DEBUG nova.network.neutron [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 561.039538] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Acquiring lock "da94a435-b36c-42b2-8c84-564265872048" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.040651] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Lock "da94a435-b36c-42b2-8c84-564265872048" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.069429] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Acquiring lock "7b159d98-9fd2-4eca-9755-cd2cac502798" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.069666] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Lock "7b159d98-9fd2-4eca-9755-cd2cac502798" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.094120] env[62385]: DEBUG nova.network.neutron [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.107224] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Acquiring lock "608c13da-6625-417d-a7b3-10821ae638d8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.107434] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Lock "608c13da-6625-417d-a7b3-10821ae638d8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.197592] env[62385]: DEBUG nova.network.neutron [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.278170] env[62385]: DEBUG nova.scheduler.client.report [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 561.298569] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6febfada-a911-4ea4-a8c8-bfab8282bfca tempest-ServersAdminNegativeTestJSON-1947249881 tempest-ServersAdminNegativeTestJSON-1947249881-project-member] Lock "b434e59a-d9ed-4aa1-a5ac-212e3d3e9c0a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.317s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 561.596296] env[62385]: INFO nova.compute.manager [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] [instance: e075eece-534b-4d97-8c82-abb960ef1abe] Took 1.06 seconds to deallocate network for instance. [ 561.703909] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] Releasing lock "refresh_cache-d0f3ba00-5eee-4ff6-8999-3a318c9f0121" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 561.703909] env[62385]: DEBUG nova.compute.manager [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 561.703909] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 561.703909] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-639ea340-58c5-4395-a54e-270316aa1cb4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.717740] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64e3db53-5025-4477-8298-e4fee3e3e321 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.744307] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d0f3ba00-5eee-4ff6-8999-3a318c9f0121 could not be found. [ 561.744540] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 561.744721] env[62385]: INFO nova.compute.manager [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Took 0.04 seconds to destroy the instance on the hypervisor. [ 561.744964] env[62385]: DEBUG oslo.service.loopingcall [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 561.745202] env[62385]: DEBUG nova.compute.manager [-] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 561.745296] env[62385]: DEBUG nova.network.neutron [-] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 561.777389] env[62385]: DEBUG nova.network.neutron [-] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 561.783480] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.557s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 561.784041] env[62385]: DEBUG nova.compute.manager [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 874d90b0-5657-467c-a934-31db63982ded] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 561.788978] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.474s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.801871] env[62385]: DEBUG nova.compute.manager [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 562.044194] env[62385]: DEBUG oslo_concurrency.lockutils [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Acquiring lock "6e0dd86b-d557-4e7a-8520-9ffbd291128b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.044644] env[62385]: DEBUG oslo_concurrency.lockutils [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Lock "6e0dd86b-d557-4e7a-8520-9ffbd291128b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.170918] env[62385]: DEBUG oslo_concurrency.lockutils [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] Acquiring lock "ea0b59fe-8dd9-4893-87ff-54706eb64479" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.171781] env[62385]: DEBUG oslo_concurrency.lockutils [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] Lock "ea0b59fe-8dd9-4893-87ff-54706eb64479" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.283815] env[62385]: DEBUG nova.network.neutron [-] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.296836] env[62385]: DEBUG nova.compute.utils [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 562.301969] env[62385]: DEBUG nova.compute.manager [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 874d90b0-5657-467c-a934-31db63982ded] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 562.302374] env[62385]: DEBUG nova.network.neutron [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 874d90b0-5657-467c-a934-31db63982ded] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 562.339215] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.389762] env[62385]: DEBUG nova.policy [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4a800e72b9e5476fa0e44203f5c7b9ca', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '37c4f2d7d8c24049b1290ec8b0ad9618', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 562.639492] env[62385]: INFO nova.scheduler.client.report [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] Deleted allocations for instance e075eece-534b-4d97-8c82-abb960ef1abe [ 562.769399] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e231006d-db5f-49b6-bd93-22f5e02e9421 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.783425] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95991b11-f8b1-4dbf-b0f4-9e5c0a2014b6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.792164] env[62385]: INFO nova.compute.manager [-] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Took 1.05 seconds to deallocate network for instance. [ 562.794590] env[62385]: DEBUG nova.compute.claims [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Aborting claim: {{(pid=62385) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 562.794832] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.835761] env[62385]: DEBUG nova.compute.manager [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 874d90b0-5657-467c-a934-31db63982ded] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 562.840905] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41c472b6-b2d7-4811-814a-a5596334635e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.848593] env[62385]: DEBUG nova.network.neutron [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 874d90b0-5657-467c-a934-31db63982ded] Successfully created port: 4f01b88c-9643-4408-b6a9-8c66cfe6f82c {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 562.851793] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3326c34a-4bda-432d-9f81-30f711223c79 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.861287] env[62385]: DEBUG oslo_concurrency.lockutils [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] Acquiring lock "7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.862189] env[62385]: DEBUG oslo_concurrency.lockutils [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] Lock "7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.874116] env[62385]: DEBUG nova.compute.provider_tree [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 563.150181] env[62385]: DEBUG oslo_concurrency.lockutils [None req-315cab5b-f00a-40c1-9585-5ff848c5b206 tempest-ServerDiagnosticsTest-1143893318 tempest-ServerDiagnosticsTest-1143893318-project-member] Lock "e075eece-534b-4d97-8c82-abb960ef1abe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.947s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 563.326971] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] Acquiring lock "1b459bac-e27d-4af5-bfc2-49898f5ced25" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.327379] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] Lock "1b459bac-e27d-4af5-bfc2-49898f5ced25" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.380262] env[62385]: DEBUG nova.scheduler.client.report [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 563.409973] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Acquiring lock "d3224596-4e38-4c9c-abd9-59a34bbf59c7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.410375] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Lock "d3224596-4e38-4c9c-abd9-59a34bbf59c7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.436647] env[62385]: DEBUG nova.compute.manager [req-7c08e639-df13-417a-9853-dbaa99b6a1e9 req-7163fe46-4d15-4c3d-aca0-fa5b057ec764 service nova] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Received event network-changed-b093b332-dcaa-435e-8b66-3b749ce492d1 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 563.436863] env[62385]: DEBUG nova.compute.manager [req-7c08e639-df13-417a-9853-dbaa99b6a1e9 req-7163fe46-4d15-4c3d-aca0-fa5b057ec764 service nova] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Refreshing instance network info cache due to event network-changed-b093b332-dcaa-435e-8b66-3b749ce492d1. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 563.437089] env[62385]: DEBUG oslo_concurrency.lockutils [req-7c08e639-df13-417a-9853-dbaa99b6a1e9 req-7163fe46-4d15-4c3d-aca0-fa5b057ec764 service nova] Acquiring lock "refresh_cache-d0f3ba00-5eee-4ff6-8999-3a318c9f0121" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 563.437232] env[62385]: DEBUG oslo_concurrency.lockutils [req-7c08e639-df13-417a-9853-dbaa99b6a1e9 req-7163fe46-4d15-4c3d-aca0-fa5b057ec764 service nova] Acquired lock "refresh_cache-d0f3ba00-5eee-4ff6-8999-3a318c9f0121" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 563.437391] env[62385]: DEBUG nova.network.neutron [req-7c08e639-df13-417a-9853-dbaa99b6a1e9 req-7163fe46-4d15-4c3d-aca0-fa5b057ec764 service nova] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Refreshing network info cache for port b093b332-dcaa-435e-8b66-3b749ce492d1 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 563.658589] env[62385]: DEBUG nova.compute.manager [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 563.850559] env[62385]: DEBUG nova.compute.manager [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 874d90b0-5657-467c-a934-31db63982ded] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 563.884014] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.095s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 563.886070] env[62385]: ERROR nova.compute.manager [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] [instance: 15703806-9f67-424d-aadb-634634810993] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2ce7db43-289a-411f-8802-33921aed5c9e, please check neutron logs for more information. [ 563.886070] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] Traceback (most recent call last): [ 563.886070] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 563.886070] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] self.driver.spawn(context, instance, image_meta, [ 563.886070] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 563.886070] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] self._vmops.spawn(context, instance, image_meta, injected_files, [ 563.886070] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 563.886070] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] vm_ref = self.build_virtual_machine(instance, [ 563.886070] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 563.886070] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] vif_infos = vmwarevif.get_vif_info(self._session, [ 563.886070] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 563.886434] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] for vif in network_info: [ 563.886434] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 563.886434] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] return self._sync_wrapper(fn, *args, **kwargs) [ 563.886434] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 563.886434] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] self.wait() [ 563.886434] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 563.886434] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] self[:] = self._gt.wait() [ 563.886434] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 563.886434] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] return self._exit_event.wait() [ 563.886434] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 563.886434] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] result = hub.switch() [ 563.886434] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 563.886434] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] return self.greenlet.switch() [ 563.886808] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 563.886808] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] result = function(*args, **kwargs) [ 563.886808] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 563.886808] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] return func(*args, **kwargs) [ 563.886808] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 563.886808] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] raise e [ 563.886808] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 563.886808] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] nwinfo = self.network_api.allocate_for_instance( [ 563.886808] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 563.886808] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] created_port_ids = self._update_ports_for_instance( [ 563.886808] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 563.886808] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] with excutils.save_and_reraise_exception(): [ 563.886808] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 563.887256] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] self.force_reraise() [ 563.887256] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 563.887256] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] raise self.value [ 563.887256] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 563.887256] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] updated_port = self._update_port( [ 563.887256] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 563.887256] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] _ensure_no_port_binding_failure(port) [ 563.887256] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 563.887256] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] raise exception.PortBindingFailed(port_id=port['id']) [ 563.887256] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] nova.exception.PortBindingFailed: Binding failed for port 2ce7db43-289a-411f-8802-33921aed5c9e, please check neutron logs for more information. [ 563.887256] env[62385]: ERROR nova.compute.manager [instance: 15703806-9f67-424d-aadb-634634810993] [ 563.887570] env[62385]: DEBUG nova.compute.utils [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] [instance: 15703806-9f67-424d-aadb-634634810993] Binding failed for port 2ce7db43-289a-411f-8802-33921aed5c9e, please check neutron logs for more information. {{(pid=62385) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 563.888998] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.046s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.891767] env[62385]: INFO nova.compute.claims [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 563.894946] env[62385]: DEBUG nova.compute.manager [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] [instance: 15703806-9f67-424d-aadb-634634810993] Build of instance 15703806-9f67-424d-aadb-634634810993 was re-scheduled: Binding failed for port 2ce7db43-289a-411f-8802-33921aed5c9e, please check neutron logs for more information. {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 563.895575] env[62385]: DEBUG nova.compute.manager [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] [instance: 15703806-9f67-424d-aadb-634634810993] Unplugging VIFs for instance {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 563.895693] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] Acquiring lock "refresh_cache-15703806-9f67-424d-aadb-634634810993" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 563.895891] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] Acquired lock "refresh_cache-15703806-9f67-424d-aadb-634634810993" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 563.896118] env[62385]: DEBUG nova.network.neutron [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] [instance: 15703806-9f67-424d-aadb-634634810993] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 563.899108] env[62385]: DEBUG nova.virt.hardware [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 563.899729] env[62385]: DEBUG nova.virt.hardware [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 563.899729] env[62385]: DEBUG nova.virt.hardware [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 563.899729] env[62385]: DEBUG nova.virt.hardware [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 563.899902] env[62385]: DEBUG nova.virt.hardware [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 563.900057] env[62385]: DEBUG nova.virt.hardware [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 563.900318] env[62385]: DEBUG nova.virt.hardware [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 563.900511] env[62385]: DEBUG nova.virt.hardware [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 563.900715] env[62385]: DEBUG nova.virt.hardware [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 563.900910] env[62385]: DEBUG nova.virt.hardware [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 563.901140] env[62385]: DEBUG nova.virt.hardware [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 563.902620] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79b7ed73-c834-4dca-9746-e28022fccb21 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.914734] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8c15b6d-324f-4e0a-bbcb-6296dc41ee1a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.943174] env[62385]: DEBUG nova.network.neutron [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] [instance: 15703806-9f67-424d-aadb-634634810993] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 563.995128] env[62385]: DEBUG nova.network.neutron [req-7c08e639-df13-417a-9853-dbaa99b6a1e9 req-7163fe46-4d15-4c3d-aca0-fa5b057ec764 service nova] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 564.146025] env[62385]: DEBUG oslo_concurrency.lockutils [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] Acquiring lock "4c96b9c6-40b9-4dea-8dd3-5e8e6372c261" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.146258] env[62385]: DEBUG oslo_concurrency.lockutils [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] Lock "4c96b9c6-40b9-4dea-8dd3-5e8e6372c261" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.189060] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.255763] env[62385]: DEBUG nova.network.neutron [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] [instance: 15703806-9f67-424d-aadb-634634810993] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.299184] env[62385]: DEBUG nova.network.neutron [req-7c08e639-df13-417a-9853-dbaa99b6a1e9 req-7163fe46-4d15-4c3d-aca0-fa5b057ec764 service nova] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.760990] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] Releasing lock "refresh_cache-15703806-9f67-424d-aadb-634634810993" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 564.762530] env[62385]: DEBUG nova.compute.manager [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 564.765150] env[62385]: DEBUG nova.compute.manager [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] [instance: 15703806-9f67-424d-aadb-634634810993] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 564.765150] env[62385]: DEBUG nova.network.neutron [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] [instance: 15703806-9f67-424d-aadb-634634810993] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 564.797205] env[62385]: DEBUG nova.network.neutron [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] [instance: 15703806-9f67-424d-aadb-634634810993] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 564.802358] env[62385]: DEBUG oslo_concurrency.lockutils [req-7c08e639-df13-417a-9853-dbaa99b6a1e9 req-7163fe46-4d15-4c3d-aca0-fa5b057ec764 service nova] Releasing lock "refresh_cache-d0f3ba00-5eee-4ff6-8999-3a318c9f0121" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 564.802600] env[62385]: DEBUG nova.compute.manager [req-7c08e639-df13-417a-9853-dbaa99b6a1e9 req-7163fe46-4d15-4c3d-aca0-fa5b057ec764 service nova] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Received event network-vif-deleted-b093b332-dcaa-435e-8b66-3b749ce492d1 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 565.081034] env[62385]: ERROR nova.compute.manager [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4f01b88c-9643-4408-b6a9-8c66cfe6f82c, please check neutron logs for more information. [ 565.081034] env[62385]: ERROR nova.compute.manager Traceback (most recent call last): [ 565.081034] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 565.081034] env[62385]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 565.081034] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 565.081034] env[62385]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 565.081034] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 565.081034] env[62385]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 565.081034] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 565.081034] env[62385]: ERROR nova.compute.manager self.force_reraise() [ 565.081034] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 565.081034] env[62385]: ERROR nova.compute.manager raise self.value [ 565.081034] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 565.081034] env[62385]: ERROR nova.compute.manager updated_port = self._update_port( [ 565.081034] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 565.081034] env[62385]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 565.081544] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 565.081544] env[62385]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 565.081544] env[62385]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4f01b88c-9643-4408-b6a9-8c66cfe6f82c, please check neutron logs for more information. [ 565.081544] env[62385]: ERROR nova.compute.manager [ 565.081544] env[62385]: Traceback (most recent call last): [ 565.081544] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 565.081544] env[62385]: listener.cb(fileno) [ 565.081544] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 565.081544] env[62385]: result = function(*args, **kwargs) [ 565.081544] env[62385]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 565.081544] env[62385]: return func(*args, **kwargs) [ 565.081544] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 565.081544] env[62385]: raise e [ 565.081544] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 565.081544] env[62385]: nwinfo = self.network_api.allocate_for_instance( [ 565.081544] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 565.081544] env[62385]: created_port_ids = self._update_ports_for_instance( [ 565.081544] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 565.081544] env[62385]: with excutils.save_and_reraise_exception(): [ 565.081544] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 565.081544] env[62385]: self.force_reraise() [ 565.081544] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 565.081544] env[62385]: raise self.value [ 565.081544] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 565.081544] env[62385]: updated_port = self._update_port( [ 565.081544] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 565.081544] env[62385]: _ensure_no_port_binding_failure(port) [ 565.081544] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 565.081544] env[62385]: raise exception.PortBindingFailed(port_id=port['id']) [ 565.082434] env[62385]: nova.exception.PortBindingFailed: Binding failed for port 4f01b88c-9643-4408-b6a9-8c66cfe6f82c, please check neutron logs for more information. [ 565.082434] env[62385]: Removing descriptor: 19 [ 565.082434] env[62385]: ERROR nova.compute.manager [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 874d90b0-5657-467c-a934-31db63982ded] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4f01b88c-9643-4408-b6a9-8c66cfe6f82c, please check neutron logs for more information. [ 565.082434] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] Traceback (most recent call last): [ 565.082434] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 565.082434] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] yield resources [ 565.082434] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 565.082434] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] self.driver.spawn(context, instance, image_meta, [ 565.082434] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 565.082434] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] self._vmops.spawn(context, instance, image_meta, injected_files, [ 565.082434] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 565.082434] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] vm_ref = self.build_virtual_machine(instance, [ 565.082812] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 565.082812] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] vif_infos = vmwarevif.get_vif_info(self._session, [ 565.082812] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 565.082812] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] for vif in network_info: [ 565.082812] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 565.082812] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] return self._sync_wrapper(fn, *args, **kwargs) [ 565.082812] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 565.082812] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] self.wait() [ 565.082812] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 565.082812] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] self[:] = self._gt.wait() [ 565.082812] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 565.082812] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] return self._exit_event.wait() [ 565.082812] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 565.083392] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] result = hub.switch() [ 565.083392] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 565.083392] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] return self.greenlet.switch() [ 565.083392] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 565.083392] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] result = function(*args, **kwargs) [ 565.083392] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 565.083392] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] return func(*args, **kwargs) [ 565.083392] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 565.083392] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] raise e [ 565.083392] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 565.083392] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] nwinfo = self.network_api.allocate_for_instance( [ 565.083392] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 565.083392] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] created_port_ids = self._update_ports_for_instance( [ 565.083846] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 565.083846] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] with excutils.save_and_reraise_exception(): [ 565.083846] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 565.083846] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] self.force_reraise() [ 565.083846] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 565.083846] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] raise self.value [ 565.083846] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 565.083846] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] updated_port = self._update_port( [ 565.083846] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 565.083846] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] _ensure_no_port_binding_failure(port) [ 565.083846] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 565.083846] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] raise exception.PortBindingFailed(port_id=port['id']) [ 565.084205] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] nova.exception.PortBindingFailed: Binding failed for port 4f01b88c-9643-4408-b6a9-8c66cfe6f82c, please check neutron logs for more information. [ 565.084205] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] [ 565.084205] env[62385]: INFO nova.compute.manager [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 874d90b0-5657-467c-a934-31db63982ded] Terminating instance [ 565.086663] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Acquiring lock "refresh_cache-874d90b0-5657-467c-a934-31db63982ded" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 565.086663] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Acquired lock "refresh_cache-874d90b0-5657-467c-a934-31db63982ded" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 565.086846] env[62385]: DEBUG nova.network.neutron [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 874d90b0-5657-467c-a934-31db63982ded] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 565.300307] env[62385]: DEBUG nova.network.neutron [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] [instance: 15703806-9f67-424d-aadb-634634810993] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 565.464834] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4df56a2a-f009-4632-b157-c6b8dfda8373 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.475022] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8fcc32c-839b-4d13-9dbe-c43a5fc1bcf8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.513515] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96e073b9-4795-4e2a-b477-f54e8869f88e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.522278] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bddd5572-0135-417a-962b-7d3bde2f39d6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.538394] env[62385]: DEBUG nova.compute.provider_tree [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 565.623050] env[62385]: DEBUG nova.network.neutron [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 874d90b0-5657-467c-a934-31db63982ded] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 565.803296] env[62385]: INFO nova.compute.manager [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] [instance: 15703806-9f67-424d-aadb-634634810993] Took 1.04 seconds to deallocate network for instance. [ 565.959572] env[62385]: DEBUG nova.compute.manager [req-34b2ed50-3367-4bf0-b23a-8f788e2067eb req-baa14851-d5ea-4611-9187-598d7e6abe6a service nova] [instance: 874d90b0-5657-467c-a934-31db63982ded] Received event network-changed-4f01b88c-9643-4408-b6a9-8c66cfe6f82c {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 565.959686] env[62385]: DEBUG nova.compute.manager [req-34b2ed50-3367-4bf0-b23a-8f788e2067eb req-baa14851-d5ea-4611-9187-598d7e6abe6a service nova] [instance: 874d90b0-5657-467c-a934-31db63982ded] Refreshing instance network info cache due to event network-changed-4f01b88c-9643-4408-b6a9-8c66cfe6f82c. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 565.959875] env[62385]: DEBUG oslo_concurrency.lockutils [req-34b2ed50-3367-4bf0-b23a-8f788e2067eb req-baa14851-d5ea-4611-9187-598d7e6abe6a service nova] Acquiring lock "refresh_cache-874d90b0-5657-467c-a934-31db63982ded" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 565.968194] env[62385]: DEBUG nova.network.neutron [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 874d90b0-5657-467c-a934-31db63982ded] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 566.040997] env[62385]: DEBUG nova.scheduler.client.report [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 566.434575] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquiring lock "866451cc-96c5-433a-a903-9faa1eed538a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.434912] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "866451cc-96c5-433a-a903-9faa1eed538a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.471137] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Releasing lock "refresh_cache-874d90b0-5657-467c-a934-31db63982ded" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 566.471543] env[62385]: DEBUG nova.compute.manager [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 874d90b0-5657-467c-a934-31db63982ded] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 566.471730] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 874d90b0-5657-467c-a934-31db63982ded] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 566.472049] env[62385]: DEBUG oslo_concurrency.lockutils [req-34b2ed50-3367-4bf0-b23a-8f788e2067eb req-baa14851-d5ea-4611-9187-598d7e6abe6a service nova] Acquired lock "refresh_cache-874d90b0-5657-467c-a934-31db63982ded" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 566.472224] env[62385]: DEBUG nova.network.neutron [req-34b2ed50-3367-4bf0-b23a-8f788e2067eb req-baa14851-d5ea-4611-9187-598d7e6abe6a service nova] [instance: 874d90b0-5657-467c-a934-31db63982ded] Refreshing network info cache for port 4f01b88c-9643-4408-b6a9-8c66cfe6f82c {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 566.473307] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d4ecd357-d6e7-41a5-a3a0-de470f41a7c9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.487296] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9deb231f-6ccc-4654-8d78-3862fed80bd7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.514379] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 874d90b0-5657-467c-a934-31db63982ded] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 874d90b0-5657-467c-a934-31db63982ded could not be found. [ 566.514613] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 874d90b0-5657-467c-a934-31db63982ded] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 566.514787] env[62385]: INFO nova.compute.manager [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 874d90b0-5657-467c-a934-31db63982ded] Took 0.04 seconds to destroy the instance on the hypervisor. [ 566.515034] env[62385]: DEBUG oslo.service.loopingcall [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 566.515258] env[62385]: DEBUG nova.compute.manager [-] [instance: 874d90b0-5657-467c-a934-31db63982ded] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 566.515345] env[62385]: DEBUG nova.network.neutron [-] [instance: 874d90b0-5657-467c-a934-31db63982ded] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 566.543528] env[62385]: DEBUG nova.network.neutron [-] [instance: 874d90b0-5657-467c-a934-31db63982ded] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 566.546785] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.657s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 566.546908] env[62385]: DEBUG nova.compute.manager [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 566.550044] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.696s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.551588] env[62385]: INFO nova.compute.claims [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 566.844923] env[62385]: INFO nova.scheduler.client.report [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] Deleted allocations for instance 15703806-9f67-424d-aadb-634634810993 [ 567.030320] env[62385]: DEBUG nova.network.neutron [req-34b2ed50-3367-4bf0-b23a-8f788e2067eb req-baa14851-d5ea-4611-9187-598d7e6abe6a service nova] [instance: 874d90b0-5657-467c-a934-31db63982ded] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 567.048143] env[62385]: DEBUG nova.network.neutron [-] [instance: 874d90b0-5657-467c-a934-31db63982ded] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.058761] env[62385]: DEBUG nova.compute.utils [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 567.061740] env[62385]: DEBUG nova.compute.manager [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 567.062485] env[62385]: DEBUG nova.network.neutron [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 567.240414] env[62385]: DEBUG nova.policy [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4a800e72b9e5476fa0e44203f5c7b9ca', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '37c4f2d7d8c24049b1290ec8b0ad9618', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 567.351758] env[62385]: DEBUG nova.network.neutron [req-34b2ed50-3367-4bf0-b23a-8f788e2067eb req-baa14851-d5ea-4611-9187-598d7e6abe6a service nova] [instance: 874d90b0-5657-467c-a934-31db63982ded] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.355535] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f5ebdbc4-f6f7-43f3-bc1e-1bff6d152156 tempest-ServersV294TestFqdnHostnames-979026060 tempest-ServersV294TestFqdnHostnames-979026060-project-member] Lock "15703806-9f67-424d-aadb-634634810993" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.997s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 567.553774] env[62385]: INFO nova.compute.manager [-] [instance: 874d90b0-5657-467c-a934-31db63982ded] Took 1.04 seconds to deallocate network for instance. [ 567.557045] env[62385]: DEBUG nova.compute.claims [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 874d90b0-5657-467c-a934-31db63982ded] Aborting claim: {{(pid=62385) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 567.557045] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.565228] env[62385]: DEBUG nova.compute.manager [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 567.854851] env[62385]: DEBUG oslo_concurrency.lockutils [req-34b2ed50-3367-4bf0-b23a-8f788e2067eb req-baa14851-d5ea-4611-9187-598d7e6abe6a service nova] Releasing lock "refresh_cache-874d90b0-5657-467c-a934-31db63982ded" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 567.856043] env[62385]: DEBUG nova.compute.manager [req-34b2ed50-3367-4bf0-b23a-8f788e2067eb req-baa14851-d5ea-4611-9187-598d7e6abe6a service nova] [instance: 874d90b0-5657-467c-a934-31db63982ded] Received event network-vif-deleted-4f01b88c-9643-4408-b6a9-8c66cfe6f82c {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 567.860413] env[62385]: DEBUG nova.compute.manager [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 567.984938] env[62385]: DEBUG nova.network.neutron [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Successfully created port: 423178e4-2acd-4345-b102-c0189917bf52 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 568.056993] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3400334b-f9d1-42f9-ad6c-d0d9ec7a3d77 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.065935] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89028e15-5454-439b-a4ca-e2c45ec90768 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.103118] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29372721-6534-45f2-83d6-87dc4eb8d864 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.111240] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-944e44e3-13f1-44de-a895-ca55bf58a6e9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.127626] env[62385]: DEBUG nova.compute.provider_tree [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 568.384696] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 568.604786] env[62385]: DEBUG nova.compute.manager [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 568.625400] env[62385]: DEBUG nova.virt.hardware [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 568.625652] env[62385]: DEBUG nova.virt.hardware [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 568.625878] env[62385]: DEBUG nova.virt.hardware [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 568.626142] env[62385]: DEBUG nova.virt.hardware [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 568.626277] env[62385]: DEBUG nova.virt.hardware [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 568.626444] env[62385]: DEBUG nova.virt.hardware [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 568.626653] env[62385]: DEBUG nova.virt.hardware [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 568.626810] env[62385]: DEBUG nova.virt.hardware [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 568.626978] env[62385]: DEBUG nova.virt.hardware [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 568.627269] env[62385]: DEBUG nova.virt.hardware [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 568.627489] env[62385]: DEBUG nova.virt.hardware [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 568.628462] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e59e15a-3ec0-49b5-a6cf-93e7f7448731 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.631714] env[62385]: DEBUG nova.scheduler.client.report [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 568.642117] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dccd8346-41c4-4a48-bb41-658d949c9d8d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.138137] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.587s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 569.138137] env[62385]: DEBUG nova.compute.manager [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 569.143820] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.431s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.151362] env[62385]: INFO nova.compute.claims [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 569.645955] env[62385]: DEBUG nova.compute.utils [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 569.647378] env[62385]: DEBUG nova.compute.manager [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 569.647553] env[62385]: DEBUG nova.network.neutron [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 569.825170] env[62385]: DEBUG nova.policy [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5069827c870c4d0fa08acc5a619a6892', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '66d888fe9de54ee1b7c2c5947816694a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 569.885309] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Acquiring lock "1ddb4d6d-3872-49db-bb40-e21721241e89" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.885825] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Lock "1ddb4d6d-3872-49db-bb40-e21721241e89" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.151797] env[62385]: DEBUG nova.compute.manager [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 570.646773] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf5b9b66-9bfd-4e7c-b2f8-fb8008278ea7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.655194] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f276c942-400b-4846-9843-33e0c2b66a52 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.693449] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5482b082-d4d0-463a-9962-dac5c63c510b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.701026] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d91481d3-8485-47f4-b25f-b89a2fd33181 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.714449] env[62385]: DEBUG nova.compute.provider_tree [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 571.193174] env[62385]: DEBUG nova.compute.manager [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 571.219927] env[62385]: DEBUG nova.virt.hardware [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 571.220184] env[62385]: DEBUG nova.virt.hardware [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 571.220551] env[62385]: DEBUG nova.virt.hardware [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 571.220551] env[62385]: DEBUG nova.virt.hardware [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 571.220830] env[62385]: DEBUG nova.virt.hardware [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 571.220830] env[62385]: DEBUG nova.virt.hardware [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 571.221065] env[62385]: DEBUG nova.virt.hardware [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 571.221134] env[62385]: DEBUG nova.virt.hardware [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 571.221327] env[62385]: DEBUG nova.virt.hardware [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 571.221586] env[62385]: DEBUG nova.virt.hardware [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 571.221657] env[62385]: DEBUG nova.virt.hardware [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 571.222460] env[62385]: DEBUG nova.scheduler.client.report [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 571.225987] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ddd2a5d-70d7-48a3-84f0-8baad649eee4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.234270] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33b1ea45-54e5-4a49-879a-b83e6a3e7593 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.423936] env[62385]: DEBUG nova.network.neutron [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Successfully created port: 13c55bc2-75d3-47de-ad87-c7c6b3cf4a31 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 571.549871] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Acquiring lock "4a4b254c-46a3-4db6-9c0f-f476ae8f2b69" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.550365] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Lock "4a4b254c-46a3-4db6-9c0f-f476ae8f2b69" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.730060] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.586s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 571.730595] env[62385]: DEBUG nova.compute.manager [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 571.734660] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.307s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.239315] env[62385]: DEBUG nova.compute.utils [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 572.249429] env[62385]: DEBUG nova.compute.manager [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 572.249603] env[62385]: DEBUG nova.network.neutron [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 572.485319] env[62385]: DEBUG nova.policy [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'eb3b765d27d149fb8d329bd7d525a0bb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1739dec97e714e29a50309dc207de27f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 572.750786] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c09c6fe7-282e-48d9-bda4-dd463145330c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.754281] env[62385]: DEBUG nova.compute.manager [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 572.761375] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-238d0b46-515b-4fc6-8956-72e0e2e6764d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.794806] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2252f52e-b83d-4e83-8974-eb9d553634fb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.802483] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d441328-ab6a-4c44-9795-468427ccc1ae {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.817067] env[62385]: DEBUG nova.compute.provider_tree [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 573.322036] env[62385]: DEBUG nova.scheduler.client.report [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 573.515780] env[62385]: ERROR nova.compute.manager [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 423178e4-2acd-4345-b102-c0189917bf52, please check neutron logs for more information. [ 573.515780] env[62385]: ERROR nova.compute.manager Traceback (most recent call last): [ 573.515780] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 573.515780] env[62385]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 573.515780] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 573.515780] env[62385]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 573.515780] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 573.515780] env[62385]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 573.515780] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.515780] env[62385]: ERROR nova.compute.manager self.force_reraise() [ 573.515780] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.515780] env[62385]: ERROR nova.compute.manager raise self.value [ 573.515780] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 573.515780] env[62385]: ERROR nova.compute.manager updated_port = self._update_port( [ 573.515780] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.515780] env[62385]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 573.516394] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.516394] env[62385]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 573.516394] env[62385]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 423178e4-2acd-4345-b102-c0189917bf52, please check neutron logs for more information. [ 573.516394] env[62385]: ERROR nova.compute.manager [ 573.516394] env[62385]: Traceback (most recent call last): [ 573.516394] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 573.516394] env[62385]: listener.cb(fileno) [ 573.516394] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 573.516394] env[62385]: result = function(*args, **kwargs) [ 573.516394] env[62385]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 573.516394] env[62385]: return func(*args, **kwargs) [ 573.516394] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 573.516394] env[62385]: raise e [ 573.516394] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 573.516394] env[62385]: nwinfo = self.network_api.allocate_for_instance( [ 573.516394] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 573.516394] env[62385]: created_port_ids = self._update_ports_for_instance( [ 573.516394] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 573.516394] env[62385]: with excutils.save_and_reraise_exception(): [ 573.516394] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.516394] env[62385]: self.force_reraise() [ 573.516394] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.516394] env[62385]: raise self.value [ 573.516394] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 573.516394] env[62385]: updated_port = self._update_port( [ 573.516394] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.516394] env[62385]: _ensure_no_port_binding_failure(port) [ 573.516394] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.516394] env[62385]: raise exception.PortBindingFailed(port_id=port['id']) [ 573.517356] env[62385]: nova.exception.PortBindingFailed: Binding failed for port 423178e4-2acd-4345-b102-c0189917bf52, please check neutron logs for more information. [ 573.517356] env[62385]: Removing descriptor: 20 [ 573.517356] env[62385]: ERROR nova.compute.manager [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 423178e4-2acd-4345-b102-c0189917bf52, please check neutron logs for more information. [ 573.517356] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Traceback (most recent call last): [ 573.517356] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 573.517356] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] yield resources [ 573.517356] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 573.517356] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] self.driver.spawn(context, instance, image_meta, [ 573.517356] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 573.517356] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] self._vmops.spawn(context, instance, image_meta, injected_files, [ 573.517356] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 573.517356] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] vm_ref = self.build_virtual_machine(instance, [ 573.517724] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 573.517724] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] vif_infos = vmwarevif.get_vif_info(self._session, [ 573.517724] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 573.517724] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] for vif in network_info: [ 573.517724] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 573.517724] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] return self._sync_wrapper(fn, *args, **kwargs) [ 573.517724] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 573.517724] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] self.wait() [ 573.517724] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 573.517724] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] self[:] = self._gt.wait() [ 573.517724] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 573.517724] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] return self._exit_event.wait() [ 573.517724] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 573.518218] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] result = hub.switch() [ 573.518218] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 573.518218] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] return self.greenlet.switch() [ 573.518218] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 573.518218] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] result = function(*args, **kwargs) [ 573.518218] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 573.518218] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] return func(*args, **kwargs) [ 573.518218] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 573.518218] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] raise e [ 573.518218] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 573.518218] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] nwinfo = self.network_api.allocate_for_instance( [ 573.518218] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 573.518218] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] created_port_ids = self._update_ports_for_instance( [ 573.518688] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 573.518688] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] with excutils.save_and_reraise_exception(): [ 573.518688] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.518688] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] self.force_reraise() [ 573.518688] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.518688] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] raise self.value [ 573.518688] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 573.518688] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] updated_port = self._update_port( [ 573.518688] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.518688] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] _ensure_no_port_binding_failure(port) [ 573.518688] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.518688] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] raise exception.PortBindingFailed(port_id=port['id']) [ 573.519197] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] nova.exception.PortBindingFailed: Binding failed for port 423178e4-2acd-4345-b102-c0189917bf52, please check neutron logs for more information. [ 573.519197] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] [ 573.519197] env[62385]: INFO nova.compute.manager [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Terminating instance [ 573.520058] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Acquiring lock "refresh_cache-2abff399-f03d-4c62-ac0b-440ba54bec29" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 573.520238] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Acquired lock "refresh_cache-2abff399-f03d-4c62-ac0b-440ba54bec29" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 573.520402] env[62385]: DEBUG nova.network.neutron [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 573.698990] env[62385]: DEBUG nova.compute.manager [req-d4597187-39d8-4776-af1e-88b4112ab2da req-eb8e1b0d-60da-4421-b4ff-e72852568868 service nova] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Received event network-changed-423178e4-2acd-4345-b102-c0189917bf52 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 573.699213] env[62385]: DEBUG nova.compute.manager [req-d4597187-39d8-4776-af1e-88b4112ab2da req-eb8e1b0d-60da-4421-b4ff-e72852568868 service nova] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Refreshing instance network info cache due to event network-changed-423178e4-2acd-4345-b102-c0189917bf52. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 573.699397] env[62385]: DEBUG oslo_concurrency.lockutils [req-d4597187-39d8-4776-af1e-88b4112ab2da req-eb8e1b0d-60da-4421-b4ff-e72852568868 service nova] Acquiring lock "refresh_cache-2abff399-f03d-4c62-ac0b-440ba54bec29" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 573.769492] env[62385]: DEBUG nova.compute.manager [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 573.801807] env[62385]: DEBUG nova.virt.hardware [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 573.802068] env[62385]: DEBUG nova.virt.hardware [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 573.802233] env[62385]: DEBUG nova.virt.hardware [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 573.802428] env[62385]: DEBUG nova.virt.hardware [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 573.802572] env[62385]: DEBUG nova.virt.hardware [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 573.802714] env[62385]: DEBUG nova.virt.hardware [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 573.802915] env[62385]: DEBUG nova.virt.hardware [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 573.804546] env[62385]: DEBUG nova.virt.hardware [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 573.805728] env[62385]: DEBUG nova.virt.hardware [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 573.805728] env[62385]: DEBUG nova.virt.hardware [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 573.805728] env[62385]: DEBUG nova.virt.hardware [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 573.806174] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf2ae815-eaf1-4ac8-bb9f-83d257367e33 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.817036] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d35de43-7dfe-4d9f-8f68-e0ec4a5d4c95 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.832099] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.097s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 573.832753] env[62385]: ERROR nova.compute.manager [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9997cf6d-ec9f-49f1-8ee9-d3684fbed8e8, please check neutron logs for more information. [ 573.832753] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Traceback (most recent call last): [ 573.832753] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 573.832753] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] self.driver.spawn(context, instance, image_meta, [ 573.832753] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 573.832753] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] self._vmops.spawn(context, instance, image_meta, injected_files, [ 573.832753] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 573.832753] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] vm_ref = self.build_virtual_machine(instance, [ 573.832753] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 573.832753] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] vif_infos = vmwarevif.get_vif_info(self._session, [ 573.832753] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 573.833203] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] for vif in network_info: [ 573.833203] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 573.833203] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] return self._sync_wrapper(fn, *args, **kwargs) [ 573.833203] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 573.833203] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] self.wait() [ 573.833203] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 573.833203] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] self[:] = self._gt.wait() [ 573.833203] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 573.833203] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] return self._exit_event.wait() [ 573.833203] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 573.833203] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] result = hub.switch() [ 573.833203] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 573.833203] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] return self.greenlet.switch() [ 573.833641] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 573.833641] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] result = function(*args, **kwargs) [ 573.833641] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 573.833641] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] return func(*args, **kwargs) [ 573.833641] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 573.833641] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] raise e [ 573.833641] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 573.833641] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] nwinfo = self.network_api.allocate_for_instance( [ 573.833641] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 573.833641] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] created_port_ids = self._update_ports_for_instance( [ 573.833641] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 573.833641] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] with excutils.save_and_reraise_exception(): [ 573.833641] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.834083] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] self.force_reraise() [ 573.834083] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.834083] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] raise self.value [ 573.834083] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 573.834083] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] updated_port = self._update_port( [ 573.834083] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.834083] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] _ensure_no_port_binding_failure(port) [ 573.834083] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.834083] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] raise exception.PortBindingFailed(port_id=port['id']) [ 573.834083] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] nova.exception.PortBindingFailed: Binding failed for port 9997cf6d-ec9f-49f1-8ee9-d3684fbed8e8, please check neutron logs for more information. [ 573.834083] env[62385]: ERROR nova.compute.manager [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] [ 573.834485] env[62385]: DEBUG nova.compute.utils [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Binding failed for port 9997cf6d-ec9f-49f1-8ee9-d3684fbed8e8, please check neutron logs for more information. {{(pid=62385) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 573.838360] env[62385]: DEBUG nova.compute.manager [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Build of instance 7ab1b11a-cf7c-4f6c-aeb2-c79681136721 was re-scheduled: Binding failed for port 9997cf6d-ec9f-49f1-8ee9-d3684fbed8e8, please check neutron logs for more information. {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 573.838493] env[62385]: DEBUG nova.compute.manager [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Unplugging VIFs for instance {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 573.838674] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Acquiring lock "refresh_cache-7ab1b11a-cf7c-4f6c-aeb2-c79681136721" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 573.838844] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Acquired lock "refresh_cache-7ab1b11a-cf7c-4f6c-aeb2-c79681136721" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 573.838939] env[62385]: DEBUG nova.network.neutron [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 573.839957] env[62385]: DEBUG oslo_concurrency.lockutils [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.293s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.841562] env[62385]: INFO nova.compute.claims [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 573.863381] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Acquiring lock "3e184125-28af-469b-83cc-4ab8859e0e2f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.863381] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Lock "3e184125-28af-469b-83cc-4ab8859e0e2f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.969354] env[62385]: DEBUG nova.network.neutron [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Successfully created port: 72a53ea0-2518-4f4e-8184-42d5b63b7147 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 574.077907] env[62385]: DEBUG nova.network.neutron [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 574.377932] env[62385]: DEBUG nova.network.neutron [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 574.406338] env[62385]: DEBUG nova.network.neutron [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.541954] env[62385]: DEBUG nova.network.neutron [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.910153] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Releasing lock "refresh_cache-2abff399-f03d-4c62-ac0b-440ba54bec29" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.911219] env[62385]: DEBUG nova.compute.manager [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 574.911219] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 574.911219] env[62385]: DEBUG oslo_concurrency.lockutils [req-d4597187-39d8-4776-af1e-88b4112ab2da req-eb8e1b0d-60da-4421-b4ff-e72852568868 service nova] Acquired lock "refresh_cache-2abff399-f03d-4c62-ac0b-440ba54bec29" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 574.912595] env[62385]: DEBUG nova.network.neutron [req-d4597187-39d8-4776-af1e-88b4112ab2da req-eb8e1b0d-60da-4421-b4ff-e72852568868 service nova] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Refreshing network info cache for port 423178e4-2acd-4345-b102-c0189917bf52 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 574.916113] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d046ee4f-f69e-44d1-8f55-c1cf296a1f2a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.928027] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88e5db4e-b418-491e-bb8f-52a005bbdb99 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.971743] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquiring lock "33387505-c576-488b-8c9c-b064fe81a7d7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.971743] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "33387505-c576-488b-8c9c-b064fe81a7d7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.971743] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2abff399-f03d-4c62-ac0b-440ba54bec29 could not be found. [ 574.971743] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 574.971743] env[62385]: INFO nova.compute.manager [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Took 0.06 seconds to destroy the instance on the hypervisor. [ 574.972243] env[62385]: DEBUG oslo.service.loopingcall [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 574.974839] env[62385]: DEBUG nova.compute.manager [-] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 574.974942] env[62385]: DEBUG nova.network.neutron [-] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 574.997437] env[62385]: DEBUG nova.network.neutron [-] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 575.044973] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Releasing lock "refresh_cache-7ab1b11a-cf7c-4f6c-aeb2-c79681136721" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 575.045259] env[62385]: DEBUG nova.compute.manager [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 575.045496] env[62385]: DEBUG nova.compute.manager [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 575.045601] env[62385]: DEBUG nova.network.neutron [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 575.072771] env[62385]: DEBUG nova.network.neutron [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 575.440481] env[62385]: DEBUG nova.network.neutron [req-d4597187-39d8-4776-af1e-88b4112ab2da req-eb8e1b0d-60da-4421-b4ff-e72852568868 service nova] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 575.498102] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f8f01e7-9ac6-4d0e-ae8f-fe98f6782984 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.501102] env[62385]: DEBUG nova.network.neutron [-] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.507164] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91aa701f-0384-4b81-89cd-5b7a25e923b5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.537977] env[62385]: DEBUG nova.network.neutron [req-d4597187-39d8-4776-af1e-88b4112ab2da req-eb8e1b0d-60da-4421-b4ff-e72852568868 service nova] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.542404] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8ec76e0-9b21-44fd-9e84-14ca77b94e01 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.550350] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87f0274d-7745-4381-b3e0-183b46362fbd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.565540] env[62385]: DEBUG nova.compute.provider_tree [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 575.577019] env[62385]: DEBUG nova.network.neutron [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.004785] env[62385]: INFO nova.compute.manager [-] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Took 1.03 seconds to deallocate network for instance. [ 576.016367] env[62385]: ERROR nova.compute.manager [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 13c55bc2-75d3-47de-ad87-c7c6b3cf4a31, please check neutron logs for more information. [ 576.016367] env[62385]: ERROR nova.compute.manager Traceback (most recent call last): [ 576.016367] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 576.016367] env[62385]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 576.016367] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 576.016367] env[62385]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 576.016367] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 576.016367] env[62385]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 576.016367] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 576.016367] env[62385]: ERROR nova.compute.manager self.force_reraise() [ 576.016367] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 576.016367] env[62385]: ERROR nova.compute.manager raise self.value [ 576.016367] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 576.016367] env[62385]: ERROR nova.compute.manager updated_port = self._update_port( [ 576.016367] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 576.016367] env[62385]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 576.016982] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 576.016982] env[62385]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 576.016982] env[62385]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 13c55bc2-75d3-47de-ad87-c7c6b3cf4a31, please check neutron logs for more information. [ 576.016982] env[62385]: ERROR nova.compute.manager [ 576.016982] env[62385]: Traceback (most recent call last): [ 576.016982] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 576.016982] env[62385]: listener.cb(fileno) [ 576.016982] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 576.016982] env[62385]: result = function(*args, **kwargs) [ 576.016982] env[62385]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 576.016982] env[62385]: return func(*args, **kwargs) [ 576.016982] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 576.016982] env[62385]: raise e [ 576.016982] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 576.016982] env[62385]: nwinfo = self.network_api.allocate_for_instance( [ 576.016982] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 576.016982] env[62385]: created_port_ids = self._update_ports_for_instance( [ 576.016982] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 576.016982] env[62385]: with excutils.save_and_reraise_exception(): [ 576.016982] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 576.016982] env[62385]: self.force_reraise() [ 576.016982] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 576.016982] env[62385]: raise self.value [ 576.016982] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 576.016982] env[62385]: updated_port = self._update_port( [ 576.016982] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 576.016982] env[62385]: _ensure_no_port_binding_failure(port) [ 576.016982] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 576.016982] env[62385]: raise exception.PortBindingFailed(port_id=port['id']) [ 576.018323] env[62385]: nova.exception.PortBindingFailed: Binding failed for port 13c55bc2-75d3-47de-ad87-c7c6b3cf4a31, please check neutron logs for more information. [ 576.018323] env[62385]: Removing descriptor: 19 [ 576.018323] env[62385]: ERROR nova.compute.manager [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 13c55bc2-75d3-47de-ad87-c7c6b3cf4a31, please check neutron logs for more information. [ 576.018323] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Traceback (most recent call last): [ 576.018323] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 576.018323] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] yield resources [ 576.018323] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 576.018323] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] self.driver.spawn(context, instance, image_meta, [ 576.018323] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 576.018323] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] self._vmops.spawn(context, instance, image_meta, injected_files, [ 576.018323] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 576.018323] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] vm_ref = self.build_virtual_machine(instance, [ 576.019094] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 576.019094] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] vif_infos = vmwarevif.get_vif_info(self._session, [ 576.019094] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 576.019094] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] for vif in network_info: [ 576.019094] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 576.019094] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] return self._sync_wrapper(fn, *args, **kwargs) [ 576.019094] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 576.019094] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] self.wait() [ 576.019094] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 576.019094] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] self[:] = self._gt.wait() [ 576.019094] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 576.019094] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] return self._exit_event.wait() [ 576.019094] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 576.020131] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] result = hub.switch() [ 576.020131] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 576.020131] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] return self.greenlet.switch() [ 576.020131] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 576.020131] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] result = function(*args, **kwargs) [ 576.020131] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 576.020131] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] return func(*args, **kwargs) [ 576.020131] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 576.020131] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] raise e [ 576.020131] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 576.020131] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] nwinfo = self.network_api.allocate_for_instance( [ 576.020131] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 576.020131] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] created_port_ids = self._update_ports_for_instance( [ 576.021180] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 576.021180] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] with excutils.save_and_reraise_exception(): [ 576.021180] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 576.021180] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] self.force_reraise() [ 576.021180] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 576.021180] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] raise self.value [ 576.021180] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 576.021180] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] updated_port = self._update_port( [ 576.021180] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 576.021180] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] _ensure_no_port_binding_failure(port) [ 576.021180] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 576.021180] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] raise exception.PortBindingFailed(port_id=port['id']) [ 576.021616] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] nova.exception.PortBindingFailed: Binding failed for port 13c55bc2-75d3-47de-ad87-c7c6b3cf4a31, please check neutron logs for more information. [ 576.021616] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] [ 576.021616] env[62385]: INFO nova.compute.manager [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Terminating instance [ 576.021616] env[62385]: DEBUG nova.compute.claims [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Aborting claim: {{(pid=62385) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 576.021616] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.025572] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Acquiring lock "refresh_cache-9ddab157-12cf-4225-bc6f-e03ef9455d79" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 576.025572] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Acquired lock "refresh_cache-9ddab157-12cf-4225-bc6f-e03ef9455d79" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 576.025572] env[62385]: DEBUG nova.network.neutron [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 576.040762] env[62385]: DEBUG oslo_concurrency.lockutils [req-d4597187-39d8-4776-af1e-88b4112ab2da req-eb8e1b0d-60da-4421-b4ff-e72852568868 service nova] Releasing lock "refresh_cache-2abff399-f03d-4c62-ac0b-440ba54bec29" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 576.070339] env[62385]: DEBUG nova.scheduler.client.report [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 576.078316] env[62385]: INFO nova.compute.manager [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 7ab1b11a-cf7c-4f6c-aeb2-c79681136721] Took 1.03 seconds to deallocate network for instance. [ 576.556523] env[62385]: DEBUG nova.network.neutron [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 576.576395] env[62385]: DEBUG oslo_concurrency.lockutils [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.736s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 576.576907] env[62385]: DEBUG nova.compute.manager [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 576.579784] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.242s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.581081] env[62385]: INFO nova.compute.claims [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 576.718917] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquiring lock "f1d33401-35f5-4d79-abb4-26dc6faa784e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.719414] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "f1d33401-35f5-4d79-abb4-26dc6faa784e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.740556] env[62385]: DEBUG nova.compute.manager [req-d088ecae-74b7-4d2e-8668-eddd7c02deba req-defecc18-4871-4fa2-b623-b3cc8415c901 service nova] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Received event network-vif-deleted-423178e4-2acd-4345-b102-c0189917bf52 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 576.740764] env[62385]: DEBUG nova.compute.manager [req-d088ecae-74b7-4d2e-8668-eddd7c02deba req-defecc18-4871-4fa2-b623-b3cc8415c901 service nova] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Received event network-changed-13c55bc2-75d3-47de-ad87-c7c6b3cf4a31 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 576.740924] env[62385]: DEBUG nova.compute.manager [req-d088ecae-74b7-4d2e-8668-eddd7c02deba req-defecc18-4871-4fa2-b623-b3cc8415c901 service nova] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Refreshing instance network info cache due to event network-changed-13c55bc2-75d3-47de-ad87-c7c6b3cf4a31. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 576.741125] env[62385]: DEBUG oslo_concurrency.lockutils [req-d088ecae-74b7-4d2e-8668-eddd7c02deba req-defecc18-4871-4fa2-b623-b3cc8415c901 service nova] Acquiring lock "refresh_cache-9ddab157-12cf-4225-bc6f-e03ef9455d79" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 576.788050] env[62385]: DEBUG nova.network.neutron [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 577.085705] env[62385]: DEBUG nova.compute.utils [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 577.089165] env[62385]: DEBUG nova.compute.manager [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 577.089218] env[62385]: DEBUG nova.network.neutron [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 577.093250] env[62385]: ERROR nova.compute.manager [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 72a53ea0-2518-4f4e-8184-42d5b63b7147, please check neutron logs for more information. [ 577.093250] env[62385]: ERROR nova.compute.manager Traceback (most recent call last): [ 577.093250] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 577.093250] env[62385]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 577.093250] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 577.093250] env[62385]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 577.093250] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 577.093250] env[62385]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 577.093250] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 577.093250] env[62385]: ERROR nova.compute.manager self.force_reraise() [ 577.093250] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 577.093250] env[62385]: ERROR nova.compute.manager raise self.value [ 577.093250] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 577.093250] env[62385]: ERROR nova.compute.manager updated_port = self._update_port( [ 577.093250] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 577.093250] env[62385]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 577.093736] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 577.093736] env[62385]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 577.093736] env[62385]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 72a53ea0-2518-4f4e-8184-42d5b63b7147, please check neutron logs for more information. [ 577.093736] env[62385]: ERROR nova.compute.manager [ 577.093736] env[62385]: Traceback (most recent call last): [ 577.093736] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 577.093736] env[62385]: listener.cb(fileno) [ 577.093736] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 577.093736] env[62385]: result = function(*args, **kwargs) [ 577.093736] env[62385]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 577.093736] env[62385]: return func(*args, **kwargs) [ 577.093736] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 577.093736] env[62385]: raise e [ 577.093736] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 577.093736] env[62385]: nwinfo = self.network_api.allocate_for_instance( [ 577.093736] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 577.093736] env[62385]: created_port_ids = self._update_ports_for_instance( [ 577.093736] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 577.093736] env[62385]: with excutils.save_and_reraise_exception(): [ 577.093736] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 577.093736] env[62385]: self.force_reraise() [ 577.093736] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 577.093736] env[62385]: raise self.value [ 577.093736] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 577.093736] env[62385]: updated_port = self._update_port( [ 577.093736] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 577.093736] env[62385]: _ensure_no_port_binding_failure(port) [ 577.093736] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 577.093736] env[62385]: raise exception.PortBindingFailed(port_id=port['id']) [ 577.094647] env[62385]: nova.exception.PortBindingFailed: Binding failed for port 72a53ea0-2518-4f4e-8184-42d5b63b7147, please check neutron logs for more information. [ 577.094647] env[62385]: Removing descriptor: 16 [ 577.101112] env[62385]: ERROR nova.compute.manager [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 72a53ea0-2518-4f4e-8184-42d5b63b7147, please check neutron logs for more information. [ 577.101112] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Traceback (most recent call last): [ 577.101112] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 577.101112] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] yield resources [ 577.101112] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 577.101112] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] self.driver.spawn(context, instance, image_meta, [ 577.101112] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 577.101112] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 577.101112] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 577.101112] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] vm_ref = self.build_virtual_machine(instance, [ 577.101112] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 577.101556] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] vif_infos = vmwarevif.get_vif_info(self._session, [ 577.101556] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 577.101556] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] for vif in network_info: [ 577.101556] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 577.101556] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] return self._sync_wrapper(fn, *args, **kwargs) [ 577.101556] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 577.101556] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] self.wait() [ 577.101556] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 577.101556] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] self[:] = self._gt.wait() [ 577.101556] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 577.101556] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] return self._exit_event.wait() [ 577.101556] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 577.101556] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] result = hub.switch() [ 577.102075] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 577.102075] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] return self.greenlet.switch() [ 577.102075] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 577.102075] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] result = function(*args, **kwargs) [ 577.102075] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 577.102075] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] return func(*args, **kwargs) [ 577.102075] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 577.102075] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] raise e [ 577.102075] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 577.102075] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] nwinfo = self.network_api.allocate_for_instance( [ 577.102075] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 577.102075] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] created_port_ids = self._update_ports_for_instance( [ 577.102075] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 577.102511] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] with excutils.save_and_reraise_exception(): [ 577.102511] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 577.102511] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] self.force_reraise() [ 577.102511] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 577.102511] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] raise self.value [ 577.102511] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 577.102511] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] updated_port = self._update_port( [ 577.102511] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 577.102511] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] _ensure_no_port_binding_failure(port) [ 577.102511] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 577.102511] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] raise exception.PortBindingFailed(port_id=port['id']) [ 577.102511] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] nova.exception.PortBindingFailed: Binding failed for port 72a53ea0-2518-4f4e-8184-42d5b63b7147, please check neutron logs for more information. [ 577.102511] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] [ 577.102947] env[62385]: INFO nova.compute.manager [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Terminating instance [ 577.107486] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] Acquiring lock "refresh_cache-759d9575-84bf-4dec-846a-e63126dae49c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 577.107640] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] Acquired lock "refresh_cache-759d9575-84bf-4dec-846a-e63126dae49c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 577.107799] env[62385]: DEBUG nova.network.neutron [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 577.130791] env[62385]: INFO nova.scheduler.client.report [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Deleted allocations for instance 7ab1b11a-cf7c-4f6c-aeb2-c79681136721 [ 577.181051] env[62385]: DEBUG nova.policy [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd37691e999db41bcae48931e03227913', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ab79759e45bb4dcb9d0f1e2cd8823aa9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 577.290240] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Releasing lock "refresh_cache-9ddab157-12cf-4225-bc6f-e03ef9455d79" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 577.290645] env[62385]: DEBUG nova.compute.manager [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 577.290828] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 577.291148] env[62385]: DEBUG oslo_concurrency.lockutils [req-d088ecae-74b7-4d2e-8668-eddd7c02deba req-defecc18-4871-4fa2-b623-b3cc8415c901 service nova] Acquired lock "refresh_cache-9ddab157-12cf-4225-bc6f-e03ef9455d79" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 577.291313] env[62385]: DEBUG nova.network.neutron [req-d088ecae-74b7-4d2e-8668-eddd7c02deba req-defecc18-4871-4fa2-b623-b3cc8415c901 service nova] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Refreshing network info cache for port 13c55bc2-75d3-47de-ad87-c7c6b3cf4a31 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 577.292353] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9cfe493f-b8d2-49dd-b003-256399963fb8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.304020] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-903146f5-70b7-44da-ad1f-66d926aaceb0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.338716] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9ddab157-12cf-4225-bc6f-e03ef9455d79 could not be found. [ 577.338940] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 577.339133] env[62385]: INFO nova.compute.manager [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Took 0.05 seconds to destroy the instance on the hypervisor. [ 577.339380] env[62385]: DEBUG oslo.service.loopingcall [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 577.342411] env[62385]: DEBUG nova.compute.manager [-] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 577.342505] env[62385]: DEBUG nova.network.neutron [-] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 577.380829] env[62385]: DEBUG nova.network.neutron [-] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 577.594343] env[62385]: DEBUG nova.compute.manager [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 577.638182] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6cd2cd3a-0796-42f2-ad3d-f42421c53517 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Lock "7ab1b11a-cf7c-4f6c-aeb2-c79681136721" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.137s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 577.644499] env[62385]: DEBUG nova.network.neutron [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 577.689541] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4284905-96e5-4bab-91be-a689eeacac78 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.698302] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bc1ba78-9f88-4c9c-bd3b-365fb72bddd6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.731283] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d20ff1c0-e739-421a-880b-5c472666dee1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.738366] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c92ade90-1e32-4081-89a9-82898cd4d96a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.752206] env[62385]: DEBUG nova.compute.provider_tree [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 577.762250] env[62385]: DEBUG nova.network.neutron [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Successfully created port: 2dddce72-7166-457f-8a3a-f4139ac7cca4 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 577.822179] env[62385]: DEBUG nova.network.neutron [req-d088ecae-74b7-4d2e-8668-eddd7c02deba req-defecc18-4871-4fa2-b623-b3cc8415c901 service nova] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 577.850990] env[62385]: DEBUG nova.network.neutron [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 577.883336] env[62385]: DEBUG nova.network.neutron [-] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 577.931648] env[62385]: DEBUG nova.network.neutron [req-d088ecae-74b7-4d2e-8668-eddd7c02deba req-defecc18-4871-4fa2-b623-b3cc8415c901 service nova] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.141843] env[62385]: DEBUG nova.compute.manager [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 578.257387] env[62385]: DEBUG nova.scheduler.client.report [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 578.358021] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] Releasing lock "refresh_cache-759d9575-84bf-4dec-846a-e63126dae49c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 578.358021] env[62385]: DEBUG nova.compute.manager [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 578.358021] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 578.358021] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9faad295-8814-4e29-895d-72651bdcac60 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.365321] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e61b037-1e8e-4602-b86c-a2e1b8c3b04c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.388537] env[62385]: INFO nova.compute.manager [-] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Took 1.05 seconds to deallocate network for instance. [ 578.389222] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 759d9575-84bf-4dec-846a-e63126dae49c could not be found. [ 578.389559] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 578.389857] env[62385]: INFO nova.compute.manager [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Took 0.03 seconds to destroy the instance on the hypervisor. [ 578.390225] env[62385]: DEBUG oslo.service.loopingcall [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 578.391953] env[62385]: DEBUG nova.compute.manager [-] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 578.392184] env[62385]: DEBUG nova.network.neutron [-] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 578.394073] env[62385]: DEBUG nova.compute.claims [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Aborting claim: {{(pid=62385) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 578.394520] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.411439] env[62385]: DEBUG nova.network.neutron [-] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 578.436325] env[62385]: DEBUG oslo_concurrency.lockutils [req-d088ecae-74b7-4d2e-8668-eddd7c02deba req-defecc18-4871-4fa2-b623-b3cc8415c901 service nova] Releasing lock "refresh_cache-9ddab157-12cf-4225-bc6f-e03ef9455d79" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 578.609549] env[62385]: DEBUG nova.compute.manager [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 578.638944] env[62385]: DEBUG nova.virt.hardware [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:17:51Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='432277714',id=24,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-1080165605',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 578.639475] env[62385]: DEBUG nova.virt.hardware [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 578.639781] env[62385]: DEBUG nova.virt.hardware [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 578.640215] env[62385]: DEBUG nova.virt.hardware [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 578.640512] env[62385]: DEBUG nova.virt.hardware [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 578.640794] env[62385]: DEBUG nova.virt.hardware [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 578.643903] env[62385]: DEBUG nova.virt.hardware [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 578.643903] env[62385]: DEBUG nova.virt.hardware [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 578.643903] env[62385]: DEBUG nova.virt.hardware [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 578.643903] env[62385]: DEBUG nova.virt.hardware [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 578.643903] env[62385]: DEBUG nova.virt.hardware [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 578.644556] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a04f4abc-5edb-4640-8360-170b49dd730b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.651698] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d91774fa-e050-4259-ba67-3ff1d10f560e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.684388] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.694683] env[62385]: DEBUG oslo_concurrency.lockutils [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Acquiring lock "f50ddafe-f5a9-422b-b0f4-46c1b111dfbe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.694998] env[62385]: DEBUG oslo_concurrency.lockutils [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Lock "f50ddafe-f5a9-422b-b0f4-46c1b111dfbe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.768781] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.189s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 578.772696] env[62385]: DEBUG nova.compute.manager [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 578.775630] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.981s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.912587] env[62385]: DEBUG nova.network.neutron [-] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 579.280762] env[62385]: DEBUG nova.compute.utils [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 579.286481] env[62385]: DEBUG nova.compute.manager [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Not allocating networking since 'none' was specified. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 579.417982] env[62385]: INFO nova.compute.manager [-] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Took 1.03 seconds to deallocate network for instance. [ 579.422665] env[62385]: DEBUG nova.compute.claims [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Aborting claim: {{(pid=62385) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 579.422877] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.649351] env[62385]: ERROR nova.compute.manager [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2dddce72-7166-457f-8a3a-f4139ac7cca4, please check neutron logs for more information. [ 579.649351] env[62385]: ERROR nova.compute.manager Traceback (most recent call last): [ 579.649351] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 579.649351] env[62385]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 579.649351] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 579.649351] env[62385]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 579.649351] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 579.649351] env[62385]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 579.649351] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 579.649351] env[62385]: ERROR nova.compute.manager self.force_reraise() [ 579.649351] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 579.649351] env[62385]: ERROR nova.compute.manager raise self.value [ 579.649351] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 579.649351] env[62385]: ERROR nova.compute.manager updated_port = self._update_port( [ 579.649351] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 579.649351] env[62385]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 579.650273] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 579.650273] env[62385]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 579.650273] env[62385]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2dddce72-7166-457f-8a3a-f4139ac7cca4, please check neutron logs for more information. [ 579.650273] env[62385]: ERROR nova.compute.manager [ 579.650273] env[62385]: Traceback (most recent call last): [ 579.650273] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 579.650273] env[62385]: listener.cb(fileno) [ 579.650273] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 579.650273] env[62385]: result = function(*args, **kwargs) [ 579.650273] env[62385]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 579.650273] env[62385]: return func(*args, **kwargs) [ 579.650273] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 579.650273] env[62385]: raise e [ 579.650273] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 579.650273] env[62385]: nwinfo = self.network_api.allocate_for_instance( [ 579.650273] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 579.650273] env[62385]: created_port_ids = self._update_ports_for_instance( [ 579.650273] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 579.650273] env[62385]: with excutils.save_and_reraise_exception(): [ 579.650273] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 579.650273] env[62385]: self.force_reraise() [ 579.650273] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 579.650273] env[62385]: raise self.value [ 579.650273] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 579.650273] env[62385]: updated_port = self._update_port( [ 579.650273] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 579.650273] env[62385]: _ensure_no_port_binding_failure(port) [ 579.650273] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 579.650273] env[62385]: raise exception.PortBindingFailed(port_id=port['id']) [ 579.652353] env[62385]: nova.exception.PortBindingFailed: Binding failed for port 2dddce72-7166-457f-8a3a-f4139ac7cca4, please check neutron logs for more information. [ 579.652353] env[62385]: Removing descriptor: 19 [ 579.652353] env[62385]: ERROR nova.compute.manager [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2dddce72-7166-457f-8a3a-f4139ac7cca4, please check neutron logs for more information. [ 579.652353] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Traceback (most recent call last): [ 579.652353] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 579.652353] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] yield resources [ 579.652353] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 579.652353] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] self.driver.spawn(context, instance, image_meta, [ 579.652353] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 579.652353] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] self._vmops.spawn(context, instance, image_meta, injected_files, [ 579.652353] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 579.652353] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] vm_ref = self.build_virtual_machine(instance, [ 579.653016] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 579.653016] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] vif_infos = vmwarevif.get_vif_info(self._session, [ 579.653016] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 579.653016] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] for vif in network_info: [ 579.653016] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 579.653016] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] return self._sync_wrapper(fn, *args, **kwargs) [ 579.653016] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 579.653016] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] self.wait() [ 579.653016] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 579.653016] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] self[:] = self._gt.wait() [ 579.653016] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 579.653016] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] return self._exit_event.wait() [ 579.653016] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 579.653648] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] result = hub.switch() [ 579.653648] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 579.653648] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] return self.greenlet.switch() [ 579.653648] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 579.653648] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] result = function(*args, **kwargs) [ 579.653648] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 579.653648] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] return func(*args, **kwargs) [ 579.653648] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 579.653648] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] raise e [ 579.653648] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 579.653648] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] nwinfo = self.network_api.allocate_for_instance( [ 579.653648] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 579.653648] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] created_port_ids = self._update_ports_for_instance( [ 579.654171] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 579.654171] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] with excutils.save_and_reraise_exception(): [ 579.654171] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 579.654171] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] self.force_reraise() [ 579.654171] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 579.654171] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] raise self.value [ 579.654171] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 579.654171] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] updated_port = self._update_port( [ 579.654171] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 579.654171] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] _ensure_no_port_binding_failure(port) [ 579.654171] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 579.654171] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] raise exception.PortBindingFailed(port_id=port['id']) [ 579.654651] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] nova.exception.PortBindingFailed: Binding failed for port 2dddce72-7166-457f-8a3a-f4139ac7cca4, please check neutron logs for more information. [ 579.654651] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] [ 579.654651] env[62385]: INFO nova.compute.manager [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Terminating instance [ 579.655237] env[62385]: DEBUG oslo_concurrency.lockutils [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] Acquiring lock "refresh_cache-be595c9e-703e-4246-ae1c-2c77b240cb19" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 579.655403] env[62385]: DEBUG oslo_concurrency.lockutils [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] Acquired lock "refresh_cache-be595c9e-703e-4246-ae1c-2c77b240cb19" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 579.655572] env[62385]: DEBUG nova.network.neutron [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 579.713595] env[62385]: DEBUG nova.compute.manager [req-68d61cea-e0a3-4f30-aeb9-330865fa6d2b req-651e6a18-cb32-4934-8537-fad3d3223d64 service nova] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Received event network-vif-deleted-13c55bc2-75d3-47de-ad87-c7c6b3cf4a31 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 579.713903] env[62385]: DEBUG nova.compute.manager [req-68d61cea-e0a3-4f30-aeb9-330865fa6d2b req-651e6a18-cb32-4934-8537-fad3d3223d64 service nova] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Received event network-changed-72a53ea0-2518-4f4e-8184-42d5b63b7147 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 579.714103] env[62385]: DEBUG nova.compute.manager [req-68d61cea-e0a3-4f30-aeb9-330865fa6d2b req-651e6a18-cb32-4934-8537-fad3d3223d64 service nova] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Refreshing instance network info cache due to event network-changed-72a53ea0-2518-4f4e-8184-42d5b63b7147. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 579.714316] env[62385]: DEBUG oslo_concurrency.lockutils [req-68d61cea-e0a3-4f30-aeb9-330865fa6d2b req-651e6a18-cb32-4934-8537-fad3d3223d64 service nova] Acquiring lock "refresh_cache-759d9575-84bf-4dec-846a-e63126dae49c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 579.714469] env[62385]: DEBUG oslo_concurrency.lockutils [req-68d61cea-e0a3-4f30-aeb9-330865fa6d2b req-651e6a18-cb32-4934-8537-fad3d3223d64 service nova] Acquired lock "refresh_cache-759d9575-84bf-4dec-846a-e63126dae49c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 579.714598] env[62385]: DEBUG nova.network.neutron [req-68d61cea-e0a3-4f30-aeb9-330865fa6d2b req-651e6a18-cb32-4934-8537-fad3d3223d64 service nova] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Refreshing network info cache for port 72a53ea0-2518-4f4e-8184-42d5b63b7147 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 579.788708] env[62385]: DEBUG nova.compute.manager [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 579.812236] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0605a876-ab43-4673-97ad-3ede6d46393c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.820185] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af3dcb59-e043-41ef-906d-bb600254994e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.851497] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4db56d52-f460-4112-8b78-00fd72949861 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.860075] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2dfda5f-fadf-45fa-aaf6-3b4e1f492211 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.873531] env[62385]: DEBUG nova.compute.provider_tree [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 580.096457] env[62385]: DEBUG oslo_concurrency.lockutils [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Acquiring lock "8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.096731] env[62385]: DEBUG oslo_concurrency.lockutils [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Lock "8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.188721] env[62385]: DEBUG nova.network.neutron [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 580.240017] env[62385]: DEBUG nova.network.neutron [req-68d61cea-e0a3-4f30-aeb9-330865fa6d2b req-651e6a18-cb32-4934-8537-fad3d3223d64 service nova] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 580.325239] env[62385]: DEBUG nova.network.neutron [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.327085] env[62385]: DEBUG nova.network.neutron [req-68d61cea-e0a3-4f30-aeb9-330865fa6d2b req-651e6a18-cb32-4934-8537-fad3d3223d64 service nova] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.376117] env[62385]: DEBUG nova.scheduler.client.report [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 580.801430] env[62385]: DEBUG nova.compute.manager [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 580.828458] env[62385]: DEBUG nova.virt.hardware [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 580.828704] env[62385]: DEBUG nova.virt.hardware [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 580.828859] env[62385]: DEBUG nova.virt.hardware [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 580.829051] env[62385]: DEBUG nova.virt.hardware [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 580.829205] env[62385]: DEBUG nova.virt.hardware [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 580.829384] env[62385]: DEBUG nova.virt.hardware [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 580.829601] env[62385]: DEBUG nova.virt.hardware [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 580.829758] env[62385]: DEBUG nova.virt.hardware [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 580.829920] env[62385]: DEBUG nova.virt.hardware [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 580.830220] env[62385]: DEBUG nova.virt.hardware [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 580.830466] env[62385]: DEBUG nova.virt.hardware [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 580.830981] env[62385]: DEBUG oslo_concurrency.lockutils [req-68d61cea-e0a3-4f30-aeb9-330865fa6d2b req-651e6a18-cb32-4934-8537-fad3d3223d64 service nova] Releasing lock "refresh_cache-759d9575-84bf-4dec-846a-e63126dae49c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 580.831277] env[62385]: DEBUG nova.compute.manager [req-68d61cea-e0a3-4f30-aeb9-330865fa6d2b req-651e6a18-cb32-4934-8537-fad3d3223d64 service nova] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Received event network-vif-deleted-72a53ea0-2518-4f4e-8184-42d5b63b7147 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 580.831799] env[62385]: DEBUG oslo_concurrency.lockutils [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] Releasing lock "refresh_cache-be595c9e-703e-4246-ae1c-2c77b240cb19" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 580.832253] env[62385]: DEBUG nova.compute.manager [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 580.832519] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 580.833406] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0572d5e7-896f-4f4f-bc56-6f3446e7af3c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.836076] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ca24a7fa-3c95-40dc-8597-042a925ca34d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.844307] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-009323f9-3790-48ea-9f89-ccf49847f0e9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.850625] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed4afe0e-488b-4257-9fcf-63f41100ea3c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.873161] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Instance VIF info [] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 580.882343] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 580.886522] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.111s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 580.887090] env[62385]: ERROR nova.compute.manager [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b093b332-dcaa-435e-8b66-3b749ce492d1, please check neutron logs for more information. [ 580.887090] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Traceback (most recent call last): [ 580.887090] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 580.887090] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] self.driver.spawn(context, instance, image_meta, [ 580.887090] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 580.887090] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] self._vmops.spawn(context, instance, image_meta, injected_files, [ 580.887090] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 580.887090] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] vm_ref = self.build_virtual_machine(instance, [ 580.887090] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 580.887090] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] vif_infos = vmwarevif.get_vif_info(self._session, [ 580.887090] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 580.887519] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] for vif in network_info: [ 580.887519] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 580.887519] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] return self._sync_wrapper(fn, *args, **kwargs) [ 580.887519] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 580.887519] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] self.wait() [ 580.887519] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 580.887519] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] self[:] = self._gt.wait() [ 580.887519] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 580.887519] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] return self._exit_event.wait() [ 580.887519] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 580.887519] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] result = hub.switch() [ 580.887519] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 580.887519] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] return self.greenlet.switch() [ 580.887917] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 580.887917] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] result = function(*args, **kwargs) [ 580.887917] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 580.887917] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] return func(*args, **kwargs) [ 580.887917] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 580.887917] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] raise e [ 580.887917] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 580.887917] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] nwinfo = self.network_api.allocate_for_instance( [ 580.887917] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 580.887917] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] created_port_ids = self._update_ports_for_instance( [ 580.887917] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 580.887917] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] with excutils.save_and_reraise_exception(): [ 580.887917] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 580.888304] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] self.force_reraise() [ 580.888304] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 580.888304] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] raise self.value [ 580.888304] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 580.888304] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] updated_port = self._update_port( [ 580.888304] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 580.888304] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] _ensure_no_port_binding_failure(port) [ 580.888304] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 580.888304] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] raise exception.PortBindingFailed(port_id=port['id']) [ 580.888304] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] nova.exception.PortBindingFailed: Binding failed for port b093b332-dcaa-435e-8b66-3b749ce492d1, please check neutron logs for more information. [ 580.888304] env[62385]: ERROR nova.compute.manager [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] [ 580.888645] env[62385]: DEBUG nova.compute.utils [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Binding failed for port b093b332-dcaa-435e-8b66-3b749ce492d1, please check neutron logs for more information. {{(pid=62385) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 580.888766] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-17943d19-9ccb-44dd-9c96-99677bbd0392 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.890625] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance be595c9e-703e-4246-ae1c-2c77b240cb19 could not be found. [ 580.890817] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 580.891007] env[62385]: INFO nova.compute.manager [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Took 0.06 seconds to destroy the instance on the hypervisor. [ 580.891248] env[62385]: DEBUG oslo.service.loopingcall [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 580.891725] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.703s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.893128] env[62385]: INFO nova.compute.claims [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 580.895391] env[62385]: DEBUG nova.compute.manager [-] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 580.895488] env[62385]: DEBUG nova.network.neutron [-] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 580.898505] env[62385]: DEBUG nova.compute.manager [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Build of instance d0f3ba00-5eee-4ff6-8999-3a318c9f0121 was re-scheduled: Binding failed for port b093b332-dcaa-435e-8b66-3b749ce492d1, please check neutron logs for more information. {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 580.898959] env[62385]: DEBUG nova.compute.manager [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Unplugging VIFs for instance {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 580.899161] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] Acquiring lock "refresh_cache-d0f3ba00-5eee-4ff6-8999-3a318c9f0121" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 580.899313] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] Acquired lock "refresh_cache-d0f3ba00-5eee-4ff6-8999-3a318c9f0121" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 580.899521] env[62385]: DEBUG nova.network.neutron [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 580.909616] env[62385]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 580.909795] env[62385]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=62385) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 580.910114] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Folder already exists: OpenStack. Parent ref: group-v4. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 580.910489] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Creating folder: Project (134f894e24074459b57b7faafdc62026). Parent ref: group-v261107. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 580.911201] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4c8278aa-ced2-420c-a22d-69465b104bb9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.918844] env[62385]: DEBUG nova.network.neutron [-] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 580.921675] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Created folder: Project (134f894e24074459b57b7faafdc62026) in parent group-v261107. [ 580.921849] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Creating folder: Instances. Parent ref: group-v261112. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 580.922131] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-97518a0f-2886-4f40-b597-3a3766d2125c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.931658] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Created folder: Instances in parent group-v261112. [ 580.931915] env[62385]: DEBUG oslo.service.loopingcall [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 580.932421] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 580.932636] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2ebb55ca-0256-40cd-b878-c40858c88f7d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.951708] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 580.951708] env[62385]: value = "task-1205435" [ 580.951708] env[62385]: _type = "Task" [ 580.951708] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 580.963684] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205435, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 581.418927] env[62385]: DEBUG nova.network.neutron [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 581.421854] env[62385]: DEBUG nova.network.neutron [-] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.462690] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205435, 'name': CreateVM_Task, 'duration_secs': 0.2468} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 581.462881] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 581.463830] env[62385]: DEBUG oslo_vmware.service [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4a42a44-b3c1-4a3d-ac16-981de6fd8cfe {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.471615] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 581.471784] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 581.472468] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 581.472706] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-460586a3-60cf-4549-8ba1-fbea3daaf1e9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.478158] env[62385]: DEBUG oslo_vmware.api [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Waiting for the task: (returnval){ [ 581.478158] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52139ff4-2090-5672-71ed-c6748ece3315" [ 581.478158] env[62385]: _type = "Task" [ 581.478158] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 581.487449] env[62385]: DEBUG oslo_vmware.api [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52139ff4-2090-5672-71ed-c6748ece3315, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 581.506341] env[62385]: DEBUG nova.network.neutron [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.908929] env[62385]: DEBUG nova.compute.manager [req-9cde393e-a40a-4315-babf-6a415f8ac888 req-5cecbe87-52cd-421e-825a-1a4e43ef83cd service nova] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Received event network-changed-2dddce72-7166-457f-8a3a-f4139ac7cca4 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 581.908929] env[62385]: DEBUG nova.compute.manager [req-9cde393e-a40a-4315-babf-6a415f8ac888 req-5cecbe87-52cd-421e-825a-1a4e43ef83cd service nova] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Refreshing instance network info cache due to event network-changed-2dddce72-7166-457f-8a3a-f4139ac7cca4. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 581.909256] env[62385]: DEBUG oslo_concurrency.lockutils [req-9cde393e-a40a-4315-babf-6a415f8ac888 req-5cecbe87-52cd-421e-825a-1a4e43ef83cd service nova] Acquiring lock "refresh_cache-be595c9e-703e-4246-ae1c-2c77b240cb19" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 581.909343] env[62385]: DEBUG oslo_concurrency.lockutils [req-9cde393e-a40a-4315-babf-6a415f8ac888 req-5cecbe87-52cd-421e-825a-1a4e43ef83cd service nova] Acquired lock "refresh_cache-be595c9e-703e-4246-ae1c-2c77b240cb19" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 581.909557] env[62385]: DEBUG nova.network.neutron [req-9cde393e-a40a-4315-babf-6a415f8ac888 req-5cecbe87-52cd-421e-825a-1a4e43ef83cd service nova] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Refreshing network info cache for port 2dddce72-7166-457f-8a3a-f4139ac7cca4 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 581.924604] env[62385]: INFO nova.compute.manager [-] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Took 1.03 seconds to deallocate network for instance. [ 581.927962] env[62385]: DEBUG nova.compute.claims [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Aborting claim: {{(pid=62385) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 581.928245] env[62385]: DEBUG oslo_concurrency.lockutils [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.990523] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 581.990812] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 581.991062] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 581.991222] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 581.991675] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 581.991927] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5558af09-b32f-45af-8758-83a426ec1d99 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.009306] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] Releasing lock "refresh_cache-d0f3ba00-5eee-4ff6-8999-3a318c9f0121" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 582.009541] env[62385]: DEBUG nova.compute.manager [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 582.009724] env[62385]: DEBUG nova.compute.manager [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 582.009892] env[62385]: DEBUG nova.network.neutron [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 582.011542] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 582.011715] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 582.014999] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba819983-16f5-4f29-8ad5-c0b6b3eea056 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.022650] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce8a4898-197e-4267-8eb6-60242cfb0775 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.026971] env[62385]: DEBUG oslo_vmware.api [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Waiting for the task: (returnval){ [ 582.026971] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]526a2f8f-7714-9710-f025-aa811351ca37" [ 582.026971] env[62385]: _type = "Task" [ 582.026971] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 582.037497] env[62385]: DEBUG oslo_vmware.api [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]526a2f8f-7714-9710-f025-aa811351ca37, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 582.038261] env[62385]: DEBUG nova.network.neutron [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 582.358780] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b343b1f7-f078-4928-99a6-436609688d9e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.366143] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-337dad99-4082-4a6f-9d45-1defab967d19 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.396999] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39bfdcda-cb4f-486e-a880-fa79fe2b0dba {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.404299] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2298aa46-1b00-4309-be6c-82bed7bf201e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.418903] env[62385]: DEBUG nova.compute.provider_tree [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 582.441115] env[62385]: DEBUG nova.network.neutron [req-9cde393e-a40a-4315-babf-6a415f8ac888 req-5cecbe87-52cd-421e-825a-1a4e43ef83cd service nova] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 582.516216] env[62385]: DEBUG nova.network.neutron [req-9cde393e-a40a-4315-babf-6a415f8ac888 req-5cecbe87-52cd-421e-825a-1a4e43ef83cd service nova] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 582.537206] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Preparing fetch location {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 582.537477] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Creating directory with path [datastore1] vmware_temp/ef702695-0404-4886-8fdc-bc1f806496d6/c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 582.537714] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d962944e-c485-4c45-9d55-9b8b237a243a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.540185] env[62385]: DEBUG nova.network.neutron [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 582.558130] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Created directory with path [datastore1] vmware_temp/ef702695-0404-4886-8fdc-bc1f806496d6/c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 582.558333] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Fetch image to [datastore1] vmware_temp/ef702695-0404-4886-8fdc-bc1f806496d6/c0abbb8d-77e7-4f22-a256-0faf0a781109/tmp-sparse.vmdk {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 582.558502] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Downloading image file data c0abbb8d-77e7-4f22-a256-0faf0a781109 to [datastore1] vmware_temp/ef702695-0404-4886-8fdc-bc1f806496d6/c0abbb8d-77e7-4f22-a256-0faf0a781109/tmp-sparse.vmdk on the data store datastore1 {{(pid=62385) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 582.559281] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7443fc68-af77-44d0-893f-96954695de64 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.566599] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99b3cb02-e477-4d26-a0af-0716efd9b098 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.575627] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af00d761-2d98-45fe-9cf2-758d236c42e8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.607323] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61028322-fceb-4fac-ba66-18059decebe2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.612911] env[62385]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-cdbfb9fe-de72-4194-b418-fbaa22e21de4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.645024] env[62385]: DEBUG nova.virt.vmwareapi.images [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Downloading image file data c0abbb8d-77e7-4f22-a256-0faf0a781109 to the data store datastore1 {{(pid=62385) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 582.697025] env[62385]: DEBUG oslo_vmware.rw_handles [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/ef702695-0404-4886-8fdc-bc1f806496d6/c0abbb8d-77e7-4f22-a256-0faf0a781109/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62385) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 582.922727] env[62385]: DEBUG nova.scheduler.client.report [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 583.019057] env[62385]: DEBUG oslo_concurrency.lockutils [req-9cde393e-a40a-4315-babf-6a415f8ac888 req-5cecbe87-52cd-421e-825a-1a4e43ef83cd service nova] Releasing lock "refresh_cache-be595c9e-703e-4246-ae1c-2c77b240cb19" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 583.019057] env[62385]: DEBUG nova.compute.manager [req-9cde393e-a40a-4315-babf-6a415f8ac888 req-5cecbe87-52cd-421e-825a-1a4e43ef83cd service nova] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Received event network-vif-deleted-2dddce72-7166-457f-8a3a-f4139ac7cca4 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 583.042983] env[62385]: INFO nova.compute.manager [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] [instance: d0f3ba00-5eee-4ff6-8999-3a318c9f0121] Took 1.03 seconds to deallocate network for instance. [ 583.335168] env[62385]: DEBUG oslo_vmware.rw_handles [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Completed reading data from the image iterator. {{(pid=62385) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 583.335398] env[62385]: DEBUG oslo_vmware.rw_handles [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Closing write handle for https://esx7c2n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/ef702695-0404-4886-8fdc-bc1f806496d6/c0abbb8d-77e7-4f22-a256-0faf0a781109/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 583.429971] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.538s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 583.430575] env[62385]: DEBUG nova.compute.manager [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 583.435612] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.879s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.472907] env[62385]: DEBUG nova.virt.vmwareapi.images [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Downloaded image file data c0abbb8d-77e7-4f22-a256-0faf0a781109 to vmware_temp/ef702695-0404-4886-8fdc-bc1f806496d6/c0abbb8d-77e7-4f22-a256-0faf0a781109/tmp-sparse.vmdk on the data store datastore1 {{(pid=62385) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 583.474931] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Caching image {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 583.475200] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Copying Virtual Disk [datastore1] vmware_temp/ef702695-0404-4886-8fdc-bc1f806496d6/c0abbb8d-77e7-4f22-a256-0faf0a781109/tmp-sparse.vmdk to [datastore1] vmware_temp/ef702695-0404-4886-8fdc-bc1f806496d6/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 583.479567] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7c6a199e-7eef-458b-8deb-d6a839bc12e8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.487878] env[62385]: DEBUG oslo_vmware.api [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Waiting for the task: (returnval){ [ 583.487878] env[62385]: value = "task-1205436" [ 583.487878] env[62385]: _type = "Task" [ 583.487878] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 583.496715] env[62385]: DEBUG oslo_vmware.api [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Task: {'id': task-1205436, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 583.943188] env[62385]: DEBUG nova.compute.utils [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 583.950660] env[62385]: DEBUG nova.compute.manager [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 583.950934] env[62385]: DEBUG nova.network.neutron [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 584.004178] env[62385]: DEBUG oslo_vmware.api [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Task: {'id': task-1205436, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 584.009217] env[62385]: DEBUG nova.policy [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '485e5b4d1ab94dfcb661d0df99c0d40e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ef9ad0e4cc924840870f50ac98ba9bd9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 584.078843] env[62385]: INFO nova.scheduler.client.report [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] Deleted allocations for instance d0f3ba00-5eee-4ff6-8999-3a318c9f0121 [ 584.354941] env[62385]: DEBUG nova.network.neutron [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Successfully created port: f46e1314-833f-428a-960e-f4c5bbc35461 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 584.449848] env[62385]: DEBUG nova.compute.manager [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 584.460021] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf87b141-917b-47d9-a286-114065d66357 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.468308] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf97b6be-a1ce-4faf-a103-8982c7fe3ed4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.505198] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed49dc44-4039-4ad1-a544-eb541cc438fb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.513239] env[62385]: DEBUG oslo_vmware.api [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Task: {'id': task-1205436, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.682783} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 584.515673] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Copied Virtual Disk [datastore1] vmware_temp/ef702695-0404-4886-8fdc-bc1f806496d6/c0abbb8d-77e7-4f22-a256-0faf0a781109/tmp-sparse.vmdk to [datastore1] vmware_temp/ef702695-0404-4886-8fdc-bc1f806496d6/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 584.516056] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Deleting the datastore file [datastore1] vmware_temp/ef702695-0404-4886-8fdc-bc1f806496d6/c0abbb8d-77e7-4f22-a256-0faf0a781109/tmp-sparse.vmdk {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 584.516455] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7e727d30-6e23-416a-9317-41a6206396b7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.519294] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c7169fd-45b8-4107-8876-1a9cab64ad65 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.535039] env[62385]: DEBUG nova.compute.provider_tree [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 584.537602] env[62385]: DEBUG oslo_vmware.api [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Waiting for the task: (returnval){ [ 584.537602] env[62385]: value = "task-1205437" [ 584.537602] env[62385]: _type = "Task" [ 584.537602] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 584.546149] env[62385]: DEBUG oslo_vmware.api [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Task: {'id': task-1205437, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 584.592212] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e9464978-a5e0-48b4-bf39-c8494787458a tempest-VolumesAssistedSnapshotsTest-1623414594 tempest-VolumesAssistedSnapshotsTest-1623414594-project-member] Lock "d0f3ba00-5eee-4ff6-8999-3a318c9f0121" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 55.521s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 585.039293] env[62385]: DEBUG nova.scheduler.client.report [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 585.054541] env[62385]: DEBUG oslo_vmware.api [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Task: {'id': task-1205437, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.021494} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 585.055147] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 585.055395] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Moving file from [datastore1] vmware_temp/ef702695-0404-4886-8fdc-bc1f806496d6/c0abbb8d-77e7-4f22-a256-0faf0a781109 to [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109. {{(pid=62385) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 585.055618] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-2379dc12-593d-4731-ac1b-2a59e1cd48d6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.067526] env[62385]: DEBUG oslo_vmware.api [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Waiting for the task: (returnval){ [ 585.067526] env[62385]: value = "task-1205438" [ 585.067526] env[62385]: _type = "Task" [ 585.067526] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 585.077165] env[62385]: DEBUG oslo_vmware.api [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Task: {'id': task-1205438, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 585.094660] env[62385]: DEBUG nova.compute.manager [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 585.320309] env[62385]: DEBUG nova.compute.manager [req-2a339527-a781-4afd-9207-b2c97f6fdcd3 req-5cfa5e32-b305-43a0-ac0e-42c0ff04fbda service nova] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Received event network-changed-f46e1314-833f-428a-960e-f4c5bbc35461 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 585.320309] env[62385]: DEBUG nova.compute.manager [req-2a339527-a781-4afd-9207-b2c97f6fdcd3 req-5cfa5e32-b305-43a0-ac0e-42c0ff04fbda service nova] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Refreshing instance network info cache due to event network-changed-f46e1314-833f-428a-960e-f4c5bbc35461. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 585.320309] env[62385]: DEBUG oslo_concurrency.lockutils [req-2a339527-a781-4afd-9207-b2c97f6fdcd3 req-5cfa5e32-b305-43a0-ac0e-42c0ff04fbda service nova] Acquiring lock "refresh_cache-0f1d83e7-bb40-4344-a344-2fc98efe8528" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 585.320309] env[62385]: DEBUG oslo_concurrency.lockutils [req-2a339527-a781-4afd-9207-b2c97f6fdcd3 req-5cfa5e32-b305-43a0-ac0e-42c0ff04fbda service nova] Acquired lock "refresh_cache-0f1d83e7-bb40-4344-a344-2fc98efe8528" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 585.320309] env[62385]: DEBUG nova.network.neutron [req-2a339527-a781-4afd-9207-b2c97f6fdcd3 req-5cfa5e32-b305-43a0-ac0e-42c0ff04fbda service nova] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Refreshing network info cache for port f46e1314-833f-428a-960e-f4c5bbc35461 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 585.459397] env[62385]: DEBUG nova.compute.manager [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 585.481737] env[62385]: ERROR nova.compute.manager [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f46e1314-833f-428a-960e-f4c5bbc35461, please check neutron logs for more information. [ 585.481737] env[62385]: ERROR nova.compute.manager Traceback (most recent call last): [ 585.481737] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 585.481737] env[62385]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 585.481737] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 585.481737] env[62385]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 585.481737] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 585.481737] env[62385]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 585.481737] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.481737] env[62385]: ERROR nova.compute.manager self.force_reraise() [ 585.481737] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.481737] env[62385]: ERROR nova.compute.manager raise self.value [ 585.481737] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 585.481737] env[62385]: ERROR nova.compute.manager updated_port = self._update_port( [ 585.481737] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.481737] env[62385]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 585.482275] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.482275] env[62385]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 585.482275] env[62385]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f46e1314-833f-428a-960e-f4c5bbc35461, please check neutron logs for more information. [ 585.482275] env[62385]: ERROR nova.compute.manager [ 585.482275] env[62385]: Traceback (most recent call last): [ 585.482275] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 585.482275] env[62385]: listener.cb(fileno) [ 585.482275] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 585.482275] env[62385]: result = function(*args, **kwargs) [ 585.482275] env[62385]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 585.482275] env[62385]: return func(*args, **kwargs) [ 585.482275] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 585.482275] env[62385]: raise e [ 585.482275] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 585.482275] env[62385]: nwinfo = self.network_api.allocate_for_instance( [ 585.482275] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 585.482275] env[62385]: created_port_ids = self._update_ports_for_instance( [ 585.482275] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 585.482275] env[62385]: with excutils.save_and_reraise_exception(): [ 585.482275] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.482275] env[62385]: self.force_reraise() [ 585.482275] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.482275] env[62385]: raise self.value [ 585.482275] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 585.482275] env[62385]: updated_port = self._update_port( [ 585.482275] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.482275] env[62385]: _ensure_no_port_binding_failure(port) [ 585.482275] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.482275] env[62385]: raise exception.PortBindingFailed(port_id=port['id']) [ 585.483176] env[62385]: nova.exception.PortBindingFailed: Binding failed for port f46e1314-833f-428a-960e-f4c5bbc35461, please check neutron logs for more information. [ 585.483176] env[62385]: Removing descriptor: 17 [ 585.497967] env[62385]: DEBUG nova.virt.hardware [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 585.498228] env[62385]: DEBUG nova.virt.hardware [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 585.502016] env[62385]: DEBUG nova.virt.hardware [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 585.502016] env[62385]: DEBUG nova.virt.hardware [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 585.502016] env[62385]: DEBUG nova.virt.hardware [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 585.502016] env[62385]: DEBUG nova.virt.hardware [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 585.502016] env[62385]: DEBUG nova.virt.hardware [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 585.502605] env[62385]: DEBUG nova.virt.hardware [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 585.502605] env[62385]: DEBUG nova.virt.hardware [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 585.502605] env[62385]: DEBUG nova.virt.hardware [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 585.502605] env[62385]: DEBUG nova.virt.hardware [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 585.503478] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f54674fb-738f-4a63-9084-16468a11a715 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.515046] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0a1f1e3-4a31-4d59-aff6-cc600bf81ee0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.530847] env[62385]: ERROR nova.compute.manager [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f46e1314-833f-428a-960e-f4c5bbc35461, please check neutron logs for more information. [ 585.530847] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Traceback (most recent call last): [ 585.530847] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 585.530847] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] yield resources [ 585.530847] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 585.530847] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] self.driver.spawn(context, instance, image_meta, [ 585.530847] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 585.530847] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] self._vmops.spawn(context, instance, image_meta, injected_files, [ 585.530847] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 585.530847] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] vm_ref = self.build_virtual_machine(instance, [ 585.530847] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 585.531321] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] vif_infos = vmwarevif.get_vif_info(self._session, [ 585.531321] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 585.531321] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] for vif in network_info: [ 585.531321] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 585.531321] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] return self._sync_wrapper(fn, *args, **kwargs) [ 585.531321] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 585.531321] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] self.wait() [ 585.531321] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 585.531321] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] self[:] = self._gt.wait() [ 585.531321] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 585.531321] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] return self._exit_event.wait() [ 585.531321] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 585.531321] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] current.throw(*self._exc) [ 585.531710] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 585.531710] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] result = function(*args, **kwargs) [ 585.531710] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 585.531710] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] return func(*args, **kwargs) [ 585.531710] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 585.531710] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] raise e [ 585.531710] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 585.531710] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] nwinfo = self.network_api.allocate_for_instance( [ 585.531710] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 585.531710] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] created_port_ids = self._update_ports_for_instance( [ 585.531710] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 585.531710] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] with excutils.save_and_reraise_exception(): [ 585.531710] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.532114] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] self.force_reraise() [ 585.532114] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.532114] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] raise self.value [ 585.532114] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 585.532114] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] updated_port = self._update_port( [ 585.532114] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.532114] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] _ensure_no_port_binding_failure(port) [ 585.532114] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.532114] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] raise exception.PortBindingFailed(port_id=port['id']) [ 585.532114] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] nova.exception.PortBindingFailed: Binding failed for port f46e1314-833f-428a-960e-f4c5bbc35461, please check neutron logs for more information. [ 585.532114] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] [ 585.532114] env[62385]: INFO nova.compute.manager [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Terminating instance [ 585.534822] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] Acquiring lock "refresh_cache-0f1d83e7-bb40-4344-a344-2fc98efe8528" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 585.549068] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.113s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 585.549874] env[62385]: ERROR nova.compute.manager [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 874d90b0-5657-467c-a934-31db63982ded] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4f01b88c-9643-4408-b6a9-8c66cfe6f82c, please check neutron logs for more information. [ 585.549874] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] Traceback (most recent call last): [ 585.549874] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 585.549874] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] self.driver.spawn(context, instance, image_meta, [ 585.549874] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 585.549874] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] self._vmops.spawn(context, instance, image_meta, injected_files, [ 585.549874] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 585.549874] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] vm_ref = self.build_virtual_machine(instance, [ 585.549874] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 585.549874] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] vif_infos = vmwarevif.get_vif_info(self._session, [ 585.549874] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 585.550259] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] for vif in network_info: [ 585.550259] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 585.550259] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] return self._sync_wrapper(fn, *args, **kwargs) [ 585.550259] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 585.550259] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] self.wait() [ 585.550259] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 585.550259] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] self[:] = self._gt.wait() [ 585.550259] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 585.550259] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] return self._exit_event.wait() [ 585.550259] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 585.550259] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] result = hub.switch() [ 585.550259] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 585.550259] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] return self.greenlet.switch() [ 585.550624] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 585.550624] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] result = function(*args, **kwargs) [ 585.550624] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 585.550624] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] return func(*args, **kwargs) [ 585.550624] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 585.550624] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] raise e [ 585.550624] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 585.550624] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] nwinfo = self.network_api.allocate_for_instance( [ 585.550624] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 585.550624] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] created_port_ids = self._update_ports_for_instance( [ 585.550624] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 585.550624] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] with excutils.save_and_reraise_exception(): [ 585.550624] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.551042] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] self.force_reraise() [ 585.551042] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.551042] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] raise self.value [ 585.551042] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 585.551042] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] updated_port = self._update_port( [ 585.551042] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.551042] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] _ensure_no_port_binding_failure(port) [ 585.551042] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.551042] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] raise exception.PortBindingFailed(port_id=port['id']) [ 585.551042] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] nova.exception.PortBindingFailed: Binding failed for port 4f01b88c-9643-4408-b6a9-8c66cfe6f82c, please check neutron logs for more information. [ 585.551042] env[62385]: ERROR nova.compute.manager [instance: 874d90b0-5657-467c-a934-31db63982ded] [ 585.551409] env[62385]: DEBUG nova.compute.utils [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 874d90b0-5657-467c-a934-31db63982ded] Binding failed for port 4f01b88c-9643-4408-b6a9-8c66cfe6f82c, please check neutron logs for more information. {{(pid=62385) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 585.551743] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.167s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.553521] env[62385]: INFO nova.compute.claims [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 585.560336] env[62385]: DEBUG nova.compute.manager [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 874d90b0-5657-467c-a934-31db63982ded] Build of instance 874d90b0-5657-467c-a934-31db63982ded was re-scheduled: Binding failed for port 4f01b88c-9643-4408-b6a9-8c66cfe6f82c, please check neutron logs for more information. {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 585.560336] env[62385]: DEBUG nova.compute.manager [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 874d90b0-5657-467c-a934-31db63982ded] Unplugging VIFs for instance {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 585.560336] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Acquiring lock "refresh_cache-874d90b0-5657-467c-a934-31db63982ded" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 585.560336] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Acquired lock "refresh_cache-874d90b0-5657-467c-a934-31db63982ded" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 585.560875] env[62385]: DEBUG nova.network.neutron [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 874d90b0-5657-467c-a934-31db63982ded] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 585.577289] env[62385]: DEBUG oslo_vmware.api [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Task: {'id': task-1205438, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.023673} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 585.577289] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] File moved {{(pid=62385) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 585.577645] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Cleaning up location [datastore1] vmware_temp/ef702695-0404-4886-8fdc-bc1f806496d6 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 585.577645] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Deleting the datastore file [datastore1] vmware_temp/ef702695-0404-4886-8fdc-bc1f806496d6 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 585.578025] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-92c2f098-76d0-4051-8954-18f8ab28b2f3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.588373] env[62385]: DEBUG oslo_vmware.api [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Waiting for the task: (returnval){ [ 585.588373] env[62385]: value = "task-1205439" [ 585.588373] env[62385]: _type = "Task" [ 585.588373] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 585.594577] env[62385]: DEBUG oslo_vmware.api [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Task: {'id': task-1205439, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 585.623502] env[62385]: DEBUG oslo_concurrency.lockutils [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.838519] env[62385]: DEBUG nova.network.neutron [req-2a339527-a781-4afd-9207-b2c97f6fdcd3 req-5cfa5e32-b305-43a0-ac0e-42c0ff04fbda service nova] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 585.912098] env[62385]: DEBUG nova.network.neutron [req-2a339527-a781-4afd-9207-b2c97f6fdcd3 req-5cfa5e32-b305-43a0-ac0e-42c0ff04fbda service nova] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.077811] env[62385]: DEBUG nova.network.neutron [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 874d90b0-5657-467c-a934-31db63982ded] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 586.095364] env[62385]: DEBUG oslo_vmware.api [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Task: {'id': task-1205439, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.032003} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 586.095649] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 586.096402] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9804dd3-10c2-492e-b316-ffa50935525f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.101570] env[62385]: DEBUG oslo_vmware.api [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Waiting for the task: (returnval){ [ 586.101570] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52257cba-885c-eb9b-60dc-8f3b9a1e1a46" [ 586.101570] env[62385]: _type = "Task" [ 586.101570] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 586.109393] env[62385]: DEBUG oslo_vmware.api [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52257cba-885c-eb9b-60dc-8f3b9a1e1a46, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 586.163021] env[62385]: DEBUG nova.network.neutron [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 874d90b0-5657-467c-a934-31db63982ded] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.415300] env[62385]: DEBUG oslo_concurrency.lockutils [req-2a339527-a781-4afd-9207-b2c97f6fdcd3 req-5cfa5e32-b305-43a0-ac0e-42c0ff04fbda service nova] Releasing lock "refresh_cache-0f1d83e7-bb40-4344-a344-2fc98efe8528" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 586.415735] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] Acquired lock "refresh_cache-0f1d83e7-bb40-4344-a344-2fc98efe8528" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 586.415932] env[62385]: DEBUG nova.network.neutron [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 586.613443] env[62385]: DEBUG oslo_vmware.api [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52257cba-885c-eb9b-60dc-8f3b9a1e1a46, 'name': SearchDatastore_Task, 'duration_secs': 0.015349} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 586.614756] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 586.614756] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] e514a72d-6b4d-46f9-935b-3535bed4afc2/e514a72d-6b4d-46f9-935b-3535bed4afc2.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 586.614756] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-50079718-37b8-447d-92ec-f5baeb576e57 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.626195] env[62385]: DEBUG oslo_vmware.api [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Waiting for the task: (returnval){ [ 586.626195] env[62385]: value = "task-1205440" [ 586.626195] env[62385]: _type = "Task" [ 586.626195] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 586.637793] env[62385]: DEBUG oslo_vmware.api [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Task: {'id': task-1205440, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 586.671743] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Releasing lock "refresh_cache-874d90b0-5657-467c-a934-31db63982ded" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 586.671743] env[62385]: DEBUG nova.compute.manager [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 586.671743] env[62385]: DEBUG nova.compute.manager [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 874d90b0-5657-467c-a934-31db63982ded] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 586.671743] env[62385]: DEBUG nova.network.neutron [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 874d90b0-5657-467c-a934-31db63982ded] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 586.689853] env[62385]: DEBUG nova.network.neutron [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 874d90b0-5657-467c-a934-31db63982ded] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 587.092573] env[62385]: DEBUG nova.network.neutron [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 587.141059] env[62385]: DEBUG oslo_vmware.api [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Task: {'id': task-1205440, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.476182} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 587.141358] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] e514a72d-6b4d-46f9-935b-3535bed4afc2/e514a72d-6b4d-46f9-935b-3535bed4afc2.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 587.141574] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 587.141820] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4196a737-7c13-4899-8774-3f196d80f304 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.153606] env[62385]: DEBUG oslo_vmware.api [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Waiting for the task: (returnval){ [ 587.153606] env[62385]: value = "task-1205441" [ 587.153606] env[62385]: _type = "Task" [ 587.153606] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 587.163603] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ef685dc-35bd-4c33-a744-b4205f3926ac {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.169047] env[62385]: DEBUG oslo_vmware.api [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Task: {'id': task-1205441, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 587.175432] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff827f27-3864-48fd-bf6c-da808eb705e9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.205964] env[62385]: DEBUG nova.network.neutron [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 874d90b0-5657-467c-a934-31db63982ded] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.207950] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6db5d9f-103f-4167-b6e9-671c61e720a2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.216085] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5900923-15c2-43ea-8f73-b451a1cb553a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.230858] env[62385]: DEBUG nova.compute.provider_tree [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 587.239667] env[62385]: DEBUG nova.network.neutron [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.437130] env[62385]: DEBUG nova.compute.manager [req-15706892-e42d-441b-9927-a70d00858e76 req-9feec054-bb99-4d7f-b328-7665061940da service nova] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Received event network-vif-deleted-f46e1314-833f-428a-960e-f4c5bbc35461 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 587.663065] env[62385]: DEBUG oslo_vmware.api [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Task: {'id': task-1205441, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060367} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 587.664195] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 587.664833] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dba8df83-36fc-48ac-bc42-7a17d320ea2e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.687047] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Reconfiguring VM instance instance-00000012 to attach disk [datastore1] e514a72d-6b4d-46f9-935b-3535bed4afc2/e514a72d-6b4d-46f9-935b-3535bed4afc2.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 587.687047] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4e843dd0-3b2b-471f-8560-b5aa32bb9ffa {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.706990] env[62385]: DEBUG oslo_vmware.api [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Waiting for the task: (returnval){ [ 587.706990] env[62385]: value = "task-1205442" [ 587.706990] env[62385]: _type = "Task" [ 587.706990] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 587.711654] env[62385]: INFO nova.compute.manager [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 874d90b0-5657-467c-a934-31db63982ded] Took 1.05 seconds to deallocate network for instance. [ 587.718020] env[62385]: DEBUG oslo_vmware.api [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Task: {'id': task-1205442, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 587.733649] env[62385]: DEBUG nova.scheduler.client.report [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 587.742735] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] Releasing lock "refresh_cache-0f1d83e7-bb40-4344-a344-2fc98efe8528" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 587.743168] env[62385]: DEBUG nova.compute.manager [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 587.743366] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 587.744101] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6798dfb5-f544-4c88-96e7-dedda1c2d1cc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.757954] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b1525c5-77b8-4d1a-bcc7-8c356e1b69ec {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.782130] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0f1d83e7-bb40-4344-a344-2fc98efe8528 could not be found. [ 587.782367] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 587.782724] env[62385]: INFO nova.compute.manager [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Took 0.04 seconds to destroy the instance on the hypervisor. [ 587.782862] env[62385]: DEBUG oslo.service.loopingcall [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 587.783460] env[62385]: DEBUG nova.compute.manager [-] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 587.783460] env[62385]: DEBUG nova.network.neutron [-] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 587.801466] env[62385]: DEBUG nova.network.neutron [-] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 588.228501] env[62385]: DEBUG oslo_vmware.api [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Task: {'id': task-1205442, 'name': ReconfigVM_Task, 'duration_secs': 0.308723} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 588.228501] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Reconfigured VM instance instance-00000012 to attach disk [datastore1] e514a72d-6b4d-46f9-935b-3535bed4afc2/e514a72d-6b4d-46f9-935b-3535bed4afc2.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 588.228501] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e3c0e87f-7537-46b2-b746-eb75fbef93c7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.233708] env[62385]: DEBUG oslo_vmware.api [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Waiting for the task: (returnval){ [ 588.233708] env[62385]: value = "task-1205443" [ 588.233708] env[62385]: _type = "Task" [ 588.233708] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 588.239239] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.687s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 588.239239] env[62385]: DEBUG nova.compute.manager [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 588.251436] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.231s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.254137] env[62385]: DEBUG oslo_vmware.api [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Task: {'id': task-1205443, 'name': Rename_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 588.304360] env[62385]: DEBUG nova.network.neutron [-] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.742636] env[62385]: DEBUG oslo_vmware.api [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Task: {'id': task-1205443, 'name': Rename_Task, 'duration_secs': 0.156356} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 588.742955] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 588.743400] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-09f6ed27-ebf9-4c2d-8800-ec945593aa69 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.750233] env[62385]: DEBUG oslo_vmware.api [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Waiting for the task: (returnval){ [ 588.750233] env[62385]: value = "task-1205444" [ 588.750233] env[62385]: _type = "Task" [ 588.750233] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 588.756776] env[62385]: DEBUG nova.compute.utils [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 588.762123] env[62385]: INFO nova.scheduler.client.report [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Deleted allocations for instance 874d90b0-5657-467c-a934-31db63982ded [ 588.776609] env[62385]: DEBUG nova.compute.manager [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 588.776609] env[62385]: DEBUG nova.network.neutron [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 588.777991] env[62385]: DEBUG oslo_vmware.api [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Task: {'id': task-1205444, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 588.807657] env[62385]: INFO nova.compute.manager [-] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Took 1.02 seconds to deallocate network for instance. [ 588.811323] env[62385]: DEBUG nova.compute.claims [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Aborting claim: {{(pid=62385) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 588.811519] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.837784] env[62385]: DEBUG nova.policy [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3621f048b2fd45f2b983afdfc160fc9b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ad8a54b359a24a97a0290e2e0b836a9b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 589.157580] env[62385]: DEBUG nova.network.neutron [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Successfully created port: 2e8c5043-214e-41e8-85e6-46fc8ecc3a4b {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 589.267817] env[62385]: DEBUG oslo_vmware.api [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Task: {'id': task-1205444, 'name': PowerOnVM_Task, 'duration_secs': 0.509178} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 589.268186] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 589.268412] env[62385]: INFO nova.compute.manager [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Took 8.47 seconds to spawn the instance on the hypervisor. [ 589.269582] env[62385]: DEBUG nova.compute.manager [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 589.270493] env[62385]: DEBUG nova.compute.manager [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 589.273525] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce687493-4955-463a-9297-04a9a097376d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.286251] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Lock "874d90b0-5657-467c-a934-31db63982ded" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 59.090s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 589.391048] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9faf5d8-376d-46e2-b604-0e534424d133 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.404024] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2afdd0a4-3a2b-448f-9987-4d6284ec773b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.442179] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e346a12-05a3-4096-9684-a9e53f9028c2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.450386] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-798e5768-7b21-4ad5-ae4c-6b7f2c11fe92 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.464520] env[62385]: DEBUG nova.compute.provider_tree [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 589.803856] env[62385]: DEBUG nova.compute.manager [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 589.807714] env[62385]: INFO nova.compute.manager [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Took 27.50 seconds to build instance. [ 589.938801] env[62385]: DEBUG oslo_concurrency.lockutils [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "caca5304-0ad7-427c-812e-de925de63f2c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.938801] env[62385]: DEBUG oslo_concurrency.lockutils [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "caca5304-0ad7-427c-812e-de925de63f2c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.967300] env[62385]: DEBUG nova.scheduler.client.report [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 590.067919] env[62385]: DEBUG nova.compute.manager [None req-cbcdfb73-f1ee-4082-958a-6439ed99b06f tempest-ServerDiagnosticsV248Test-929408931 tempest-ServerDiagnosticsV248Test-929408931-project-admin] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 590.069488] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3efcb8b2-dc3b-4bfc-8f13-4127b8b937aa {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.077122] env[62385]: INFO nova.compute.manager [None req-cbcdfb73-f1ee-4082-958a-6439ed99b06f tempest-ServerDiagnosticsV248Test-929408931 tempest-ServerDiagnosticsV248Test-929408931-project-admin] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Retrieving diagnostics [ 590.078485] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cd17e76-0cd3-49e6-9439-6fa0b345d7e5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.162014] env[62385]: ERROR nova.compute.manager [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2e8c5043-214e-41e8-85e6-46fc8ecc3a4b, please check neutron logs for more information. [ 590.162014] env[62385]: ERROR nova.compute.manager Traceback (most recent call last): [ 590.162014] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 590.162014] env[62385]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 590.162014] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 590.162014] env[62385]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 590.162014] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 590.162014] env[62385]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 590.162014] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 590.162014] env[62385]: ERROR nova.compute.manager self.force_reraise() [ 590.162014] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 590.162014] env[62385]: ERROR nova.compute.manager raise self.value [ 590.162014] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 590.162014] env[62385]: ERROR nova.compute.manager updated_port = self._update_port( [ 590.162014] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 590.162014] env[62385]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 590.162567] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 590.162567] env[62385]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 590.162567] env[62385]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2e8c5043-214e-41e8-85e6-46fc8ecc3a4b, please check neutron logs for more information. [ 590.162567] env[62385]: ERROR nova.compute.manager [ 590.162567] env[62385]: Traceback (most recent call last): [ 590.162567] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 590.162567] env[62385]: listener.cb(fileno) [ 590.162567] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 590.162567] env[62385]: result = function(*args, **kwargs) [ 590.162567] env[62385]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 590.162567] env[62385]: return func(*args, **kwargs) [ 590.162567] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 590.162567] env[62385]: raise e [ 590.162567] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 590.162567] env[62385]: nwinfo = self.network_api.allocate_for_instance( [ 590.162567] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 590.162567] env[62385]: created_port_ids = self._update_ports_for_instance( [ 590.162567] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 590.162567] env[62385]: with excutils.save_and_reraise_exception(): [ 590.162567] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 590.162567] env[62385]: self.force_reraise() [ 590.162567] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 590.162567] env[62385]: raise self.value [ 590.162567] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 590.162567] env[62385]: updated_port = self._update_port( [ 590.162567] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 590.162567] env[62385]: _ensure_no_port_binding_failure(port) [ 590.162567] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 590.162567] env[62385]: raise exception.PortBindingFailed(port_id=port['id']) [ 590.163355] env[62385]: nova.exception.PortBindingFailed: Binding failed for port 2e8c5043-214e-41e8-85e6-46fc8ecc3a4b, please check neutron logs for more information. [ 590.163355] env[62385]: Removing descriptor: 17 [ 590.290923] env[62385]: DEBUG nova.compute.manager [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 590.313084] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bf0ae2fe-5b08-43c3-bf52-8095f4cb6987 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Lock "e514a72d-6b4d-46f9-935b-3535bed4afc2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.996s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 590.325953] env[62385]: DEBUG nova.virt.hardware [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 590.326383] env[62385]: DEBUG nova.virt.hardware [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 590.326633] env[62385]: DEBUG nova.virt.hardware [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 590.326918] env[62385]: DEBUG nova.virt.hardware [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 590.327335] env[62385]: DEBUG nova.virt.hardware [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 590.327633] env[62385]: DEBUG nova.virt.hardware [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 590.328542] env[62385]: DEBUG nova.virt.hardware [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 590.328542] env[62385]: DEBUG nova.virt.hardware [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 590.328542] env[62385]: DEBUG nova.virt.hardware [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 590.328542] env[62385]: DEBUG nova.virt.hardware [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 590.328818] env[62385]: DEBUG nova.virt.hardware [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 590.329877] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5922290-9775-42d7-b804-897ca9ac0bf7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.336957] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.338010] env[62385]: DEBUG nova.compute.manager [req-52abe7cc-3c29-4c22-afae-afb028b48f7c req-4896041e-41ff-40f5-ae52-0397bd9947fc service nova] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Received event network-changed-2e8c5043-214e-41e8-85e6-46fc8ecc3a4b {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 590.338319] env[62385]: DEBUG nova.compute.manager [req-52abe7cc-3c29-4c22-afae-afb028b48f7c req-4896041e-41ff-40f5-ae52-0397bd9947fc service nova] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Refreshing instance network info cache due to event network-changed-2e8c5043-214e-41e8-85e6-46fc8ecc3a4b. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 590.338595] env[62385]: DEBUG oslo_concurrency.lockutils [req-52abe7cc-3c29-4c22-afae-afb028b48f7c req-4896041e-41ff-40f5-ae52-0397bd9947fc service nova] Acquiring lock "refresh_cache-0a21307b-74e7-4c5d-95a3-0800915c51ff" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 590.338778] env[62385]: DEBUG oslo_concurrency.lockutils [req-52abe7cc-3c29-4c22-afae-afb028b48f7c req-4896041e-41ff-40f5-ae52-0397bd9947fc service nova] Acquired lock "refresh_cache-0a21307b-74e7-4c5d-95a3-0800915c51ff" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 590.339021] env[62385]: DEBUG nova.network.neutron [req-52abe7cc-3c29-4c22-afae-afb028b48f7c req-4896041e-41ff-40f5-ae52-0397bd9947fc service nova] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Refreshing network info cache for port 2e8c5043-214e-41e8-85e6-46fc8ecc3a4b {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 590.351999] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2f485ac-6788-4626-b7cc-0b15eb142897 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.370129] env[62385]: ERROR nova.compute.manager [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2e8c5043-214e-41e8-85e6-46fc8ecc3a4b, please check neutron logs for more information. [ 590.370129] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Traceback (most recent call last): [ 590.370129] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 590.370129] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] yield resources [ 590.370129] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 590.370129] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] self.driver.spawn(context, instance, image_meta, [ 590.370129] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 590.370129] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] self._vmops.spawn(context, instance, image_meta, injected_files, [ 590.370129] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 590.370129] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] vm_ref = self.build_virtual_machine(instance, [ 590.370129] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 590.370483] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] vif_infos = vmwarevif.get_vif_info(self._session, [ 590.370483] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 590.370483] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] for vif in network_info: [ 590.370483] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 590.370483] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] return self._sync_wrapper(fn, *args, **kwargs) [ 590.370483] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 590.370483] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] self.wait() [ 590.370483] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 590.370483] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] self[:] = self._gt.wait() [ 590.370483] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 590.370483] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] return self._exit_event.wait() [ 590.370483] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 590.370483] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] current.throw(*self._exc) [ 590.371094] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 590.371094] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] result = function(*args, **kwargs) [ 590.371094] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 590.371094] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] return func(*args, **kwargs) [ 590.371094] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 590.371094] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] raise e [ 590.371094] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 590.371094] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] nwinfo = self.network_api.allocate_for_instance( [ 590.371094] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 590.371094] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] created_port_ids = self._update_ports_for_instance( [ 590.371094] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 590.371094] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] with excutils.save_and_reraise_exception(): [ 590.371094] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 590.371377] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] self.force_reraise() [ 590.371377] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 590.371377] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] raise self.value [ 590.371377] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 590.371377] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] updated_port = self._update_port( [ 590.371377] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 590.371377] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] _ensure_no_port_binding_failure(port) [ 590.371377] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 590.371377] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] raise exception.PortBindingFailed(port_id=port['id']) [ 590.371377] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] nova.exception.PortBindingFailed: Binding failed for port 2e8c5043-214e-41e8-85e6-46fc8ecc3a4b, please check neutron logs for more information. [ 590.371377] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] [ 590.371377] env[62385]: INFO nova.compute.manager [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Terminating instance [ 590.376296] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] Acquiring lock "refresh_cache-0a21307b-74e7-4c5d-95a3-0800915c51ff" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 590.473110] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.222s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 590.474054] env[62385]: ERROR nova.compute.manager [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 423178e4-2acd-4345-b102-c0189917bf52, please check neutron logs for more information. [ 590.474054] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Traceback (most recent call last): [ 590.474054] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 590.474054] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] self.driver.spawn(context, instance, image_meta, [ 590.474054] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 590.474054] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] self._vmops.spawn(context, instance, image_meta, injected_files, [ 590.474054] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 590.474054] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] vm_ref = self.build_virtual_machine(instance, [ 590.474054] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 590.474054] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] vif_infos = vmwarevif.get_vif_info(self._session, [ 590.474054] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 590.474342] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] for vif in network_info: [ 590.474342] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 590.474342] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] return self._sync_wrapper(fn, *args, **kwargs) [ 590.474342] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 590.474342] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] self.wait() [ 590.474342] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 590.474342] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] self[:] = self._gt.wait() [ 590.474342] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 590.474342] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] return self._exit_event.wait() [ 590.474342] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 590.474342] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] result = hub.switch() [ 590.474342] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 590.474342] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] return self.greenlet.switch() [ 590.474626] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 590.474626] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] result = function(*args, **kwargs) [ 590.474626] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 590.474626] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] return func(*args, **kwargs) [ 590.474626] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 590.474626] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] raise e [ 590.474626] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 590.474626] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] nwinfo = self.network_api.allocate_for_instance( [ 590.474626] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 590.474626] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] created_port_ids = self._update_ports_for_instance( [ 590.474626] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 590.474626] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] with excutils.save_and_reraise_exception(): [ 590.474626] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 590.474920] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] self.force_reraise() [ 590.474920] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 590.474920] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] raise self.value [ 590.474920] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 590.474920] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] updated_port = self._update_port( [ 590.474920] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 590.474920] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] _ensure_no_port_binding_failure(port) [ 590.474920] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 590.474920] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] raise exception.PortBindingFailed(port_id=port['id']) [ 590.474920] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] nova.exception.PortBindingFailed: Binding failed for port 423178e4-2acd-4345-b102-c0189917bf52, please check neutron logs for more information. [ 590.474920] env[62385]: ERROR nova.compute.manager [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] [ 590.476216] env[62385]: DEBUG nova.compute.utils [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Binding failed for port 423178e4-2acd-4345-b102-c0189917bf52, please check neutron logs for more information. {{(pid=62385) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 590.477511] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.083s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.484022] env[62385]: DEBUG nova.compute.manager [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Build of instance 2abff399-f03d-4c62-ac0b-440ba54bec29 was re-scheduled: Binding failed for port 423178e4-2acd-4345-b102-c0189917bf52, please check neutron logs for more information. {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 590.484022] env[62385]: DEBUG nova.compute.manager [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Unplugging VIFs for instance {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 590.484022] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Acquiring lock "refresh_cache-2abff399-f03d-4c62-ac0b-440ba54bec29" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 590.484022] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Acquired lock "refresh_cache-2abff399-f03d-4c62-ac0b-440ba54bec29" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 590.484381] env[62385]: DEBUG nova.network.neutron [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 590.820017] env[62385]: DEBUG nova.compute.manager [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 590.861207] env[62385]: DEBUG nova.network.neutron [req-52abe7cc-3c29-4c22-afae-afb028b48f7c req-4896041e-41ff-40f5-ae52-0397bd9947fc service nova] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 590.950644] env[62385]: DEBUG nova.network.neutron [req-52abe7cc-3c29-4c22-afae-afb028b48f7c req-4896041e-41ff-40f5-ae52-0397bd9947fc service nova] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.012507] env[62385]: DEBUG nova.network.neutron [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 591.147162] env[62385]: DEBUG nova.network.neutron [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.341734] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.457194] env[62385]: DEBUG oslo_concurrency.lockutils [req-52abe7cc-3c29-4c22-afae-afb028b48f7c req-4896041e-41ff-40f5-ae52-0397bd9947fc service nova] Releasing lock "refresh_cache-0a21307b-74e7-4c5d-95a3-0800915c51ff" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 591.458288] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] Acquired lock "refresh_cache-0a21307b-74e7-4c5d-95a3-0800915c51ff" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 591.458288] env[62385]: DEBUG nova.network.neutron [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 591.535593] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-826f980c-5726-4641-aec5-03d049465e5d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.547465] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d61d08c-275f-4138-beaf-d0f8395f255a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.576800] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa142006-9057-424b-8249-9237772b8801 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.584524] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e05456b-bbec-4029-9e60-f163c4b60dec {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.597599] env[62385]: DEBUG nova.compute.provider_tree [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 591.649789] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Releasing lock "refresh_cache-2abff399-f03d-4c62-ac0b-440ba54bec29" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 591.650048] env[62385]: DEBUG nova.compute.manager [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 591.650240] env[62385]: DEBUG nova.compute.manager [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 591.650409] env[62385]: DEBUG nova.network.neutron [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 591.670505] env[62385]: DEBUG nova.network.neutron [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 591.984833] env[62385]: DEBUG nova.network.neutron [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 592.100773] env[62385]: DEBUG nova.scheduler.client.report [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 592.109568] env[62385]: DEBUG nova.network.neutron [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.173254] env[62385]: DEBUG nova.network.neutron [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.368266] env[62385]: DEBUG nova.compute.manager [req-748760c2-ae46-4b52-954e-91b3986d3d2d req-892b193c-9026-4ac5-bff4-fcbde93725fb service nova] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Received event network-vif-deleted-2e8c5043-214e-41e8-85e6-46fc8ecc3a4b {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 592.609186] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.131s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 592.609401] env[62385]: ERROR nova.compute.manager [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 13c55bc2-75d3-47de-ad87-c7c6b3cf4a31, please check neutron logs for more information. [ 592.609401] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Traceback (most recent call last): [ 592.609401] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 592.609401] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] self.driver.spawn(context, instance, image_meta, [ 592.609401] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 592.609401] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] self._vmops.spawn(context, instance, image_meta, injected_files, [ 592.609401] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 592.609401] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] vm_ref = self.build_virtual_machine(instance, [ 592.609401] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 592.609401] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] vif_infos = vmwarevif.get_vif_info(self._session, [ 592.609401] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 592.609795] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] for vif in network_info: [ 592.609795] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 592.609795] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] return self._sync_wrapper(fn, *args, **kwargs) [ 592.609795] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 592.609795] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] self.wait() [ 592.609795] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 592.609795] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] self[:] = self._gt.wait() [ 592.609795] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 592.609795] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] return self._exit_event.wait() [ 592.609795] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 592.609795] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] result = hub.switch() [ 592.609795] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 592.609795] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] return self.greenlet.switch() [ 592.610073] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 592.610073] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] result = function(*args, **kwargs) [ 592.610073] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 592.610073] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] return func(*args, **kwargs) [ 592.610073] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 592.610073] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] raise e [ 592.610073] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 592.610073] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] nwinfo = self.network_api.allocate_for_instance( [ 592.610073] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 592.610073] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] created_port_ids = self._update_ports_for_instance( [ 592.610073] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 592.610073] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] with excutils.save_and_reraise_exception(): [ 592.610073] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 592.610339] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] self.force_reraise() [ 592.610339] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 592.610339] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] raise self.value [ 592.610339] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 592.610339] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] updated_port = self._update_port( [ 592.610339] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 592.610339] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] _ensure_no_port_binding_failure(port) [ 592.610339] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 592.610339] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] raise exception.PortBindingFailed(port_id=port['id']) [ 592.610339] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] nova.exception.PortBindingFailed: Binding failed for port 13c55bc2-75d3-47de-ad87-c7c6b3cf4a31, please check neutron logs for more information. [ 592.610339] env[62385]: ERROR nova.compute.manager [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] [ 592.610581] env[62385]: DEBUG nova.compute.utils [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Binding failed for port 13c55bc2-75d3-47de-ad87-c7c6b3cf4a31, please check neutron logs for more information. {{(pid=62385) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 592.611818] env[62385]: DEBUG nova.compute.manager [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Build of instance 9ddab157-12cf-4225-bc6f-e03ef9455d79 was re-scheduled: Binding failed for port 13c55bc2-75d3-47de-ad87-c7c6b3cf4a31, please check neutron logs for more information. {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 592.612239] env[62385]: DEBUG nova.compute.manager [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Unplugging VIFs for instance {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 592.613874] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Acquiring lock "refresh_cache-9ddab157-12cf-4225-bc6f-e03ef9455d79" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 592.613874] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Acquired lock "refresh_cache-9ddab157-12cf-4225-bc6f-e03ef9455d79" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 592.613874] env[62385]: DEBUG nova.network.neutron [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 592.614647] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.930s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.618199] env[62385]: INFO nova.compute.claims [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 592.621007] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] Releasing lock "refresh_cache-0a21307b-74e7-4c5d-95a3-0800915c51ff" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 592.621383] env[62385]: DEBUG nova.compute.manager [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 592.621565] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 592.622612] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-37c49527-3b1d-462d-be42-21c37d35d8b9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.633290] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd9a7d85-f9ef-4839-9b19-f5f1db0aa411 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.658211] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0a21307b-74e7-4c5d-95a3-0800915c51ff could not be found. [ 592.658872] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 592.658872] env[62385]: INFO nova.compute.manager [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Took 0.04 seconds to destroy the instance on the hypervisor. [ 592.658872] env[62385]: DEBUG oslo.service.loopingcall [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 592.659622] env[62385]: DEBUG nova.compute.manager [-] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 592.659713] env[62385]: DEBUG nova.network.neutron [-] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 592.675109] env[62385]: DEBUG nova.network.neutron [-] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 592.676665] env[62385]: INFO nova.compute.manager [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 2abff399-f03d-4c62-ac0b-440ba54bec29] Took 1.03 seconds to deallocate network for instance. [ 593.142190] env[62385]: DEBUG nova.network.neutron [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 593.179805] env[62385]: DEBUG nova.network.neutron [-] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.234337] env[62385]: DEBUG nova.network.neutron [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.689195] env[62385]: INFO nova.compute.manager [-] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Took 1.03 seconds to deallocate network for instance. [ 593.695351] env[62385]: DEBUG nova.compute.claims [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Aborting claim: {{(pid=62385) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 593.695351] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.707185] env[62385]: INFO nova.scheduler.client.report [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Deleted allocations for instance 2abff399-f03d-4c62-ac0b-440ba54bec29 [ 593.737084] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Releasing lock "refresh_cache-9ddab157-12cf-4225-bc6f-e03ef9455d79" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 593.737356] env[62385]: DEBUG nova.compute.manager [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 593.737555] env[62385]: DEBUG nova.compute.manager [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 593.737729] env[62385]: DEBUG nova.network.neutron [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 593.766475] env[62385]: DEBUG nova.network.neutron [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 594.055153] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f6ac547-2c1d-4e0b-a6ba-8dfb35146da6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.063060] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d81bf2d7-9bb4-4145-8985-087f1dc89847 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.096503] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce312eba-1793-421d-948c-2caf9dcb4dd9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.101383] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fe8d800-dbd7-4851-b0b3-82bb92046f2c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.115280] env[62385]: DEBUG nova.compute.provider_tree [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 594.215271] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e32e1960-3e9a-41d6-8bf3-167e1e142a6e tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Lock "2abff399-f03d-4c62-ac0b-440ba54bec29" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 63.997s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 594.270258] env[62385]: DEBUG nova.network.neutron [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.618330] env[62385]: DEBUG nova.scheduler.client.report [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 594.718372] env[62385]: DEBUG nova.compute.manager [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 594.774043] env[62385]: INFO nova.compute.manager [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] [instance: 9ddab157-12cf-4225-bc6f-e03ef9455d79] Took 1.03 seconds to deallocate network for instance. [ 595.125035] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.511s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.125522] env[62385]: DEBUG nova.compute.manager [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 595.128138] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.705s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.243848] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.637598] env[62385]: DEBUG nova.compute.utils [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 595.642493] env[62385]: DEBUG nova.compute.manager [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 595.642661] env[62385]: DEBUG nova.network.neutron [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 595.686915] env[62385]: DEBUG nova.policy [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5e83312d226945c99d05fcc7f74c0978', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bf8e3c942d3445919cfbe988cca84e90', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 595.811070] env[62385]: INFO nova.scheduler.client.report [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Deleted allocations for instance 9ddab157-12cf-4225-bc6f-e03ef9455d79 [ 596.032370] env[62385]: DEBUG nova.network.neutron [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Successfully created port: e057457e-a781-4fd0-b8da-a1ea032ef872 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 596.122041] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Acquiring lock "e97ff020-61f3-4947-bb82-5c039ad17747" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.122353] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Lock "e97ff020-61f3-4947-bb82-5c039ad17747" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.143154] env[62385]: DEBUG nova.compute.manager [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 596.174399] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Acquiring lock "749e0976-2e2b-4764-865c-2e630f2edbd1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.174632] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Lock "749e0976-2e2b-4764-865c-2e630f2edbd1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.244806] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b918ae22-a525-46d1-8460-1628b0305666 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.253257] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1680eed9-d151-4e54-adec-bb14c75f35cd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.282753] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b33efdaf-672e-405b-b9bd-6b5da42fcebb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.290493] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89d4f173-18df-4529-8b60-930947ad3e58 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.303985] env[62385]: DEBUG nova.compute.provider_tree [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 596.326784] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c7998976-5b79-4054-9ac6-8d870bea5d00 tempest-ServerRescueNegativeTestJSON-1015148898 tempest-ServerRescueNegativeTestJSON-1015148898-project-member] Lock "9ddab157-12cf-4225-bc6f-e03ef9455d79" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 64.396s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 596.807926] env[62385]: DEBUG nova.scheduler.client.report [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 596.828975] env[62385]: DEBUG nova.compute.manager [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 596.874549] env[62385]: DEBUG nova.compute.manager [req-68463ccc-a535-4918-8838-4d077262c343 req-61ab5d98-b5a0-4aad-81d4-e603f8ef7e5f service nova] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Received event network-changed-e057457e-a781-4fd0-b8da-a1ea032ef872 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 596.874959] env[62385]: DEBUG nova.compute.manager [req-68463ccc-a535-4918-8838-4d077262c343 req-61ab5d98-b5a0-4aad-81d4-e603f8ef7e5f service nova] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Refreshing instance network info cache due to event network-changed-e057457e-a781-4fd0-b8da-a1ea032ef872. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 596.875250] env[62385]: DEBUG oslo_concurrency.lockutils [req-68463ccc-a535-4918-8838-4d077262c343 req-61ab5d98-b5a0-4aad-81d4-e603f8ef7e5f service nova] Acquiring lock "refresh_cache-ce6b8f31-afee-4062-b860-d054c4d37be0" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 596.875521] env[62385]: DEBUG oslo_concurrency.lockutils [req-68463ccc-a535-4918-8838-4d077262c343 req-61ab5d98-b5a0-4aad-81d4-e603f8ef7e5f service nova] Acquired lock "refresh_cache-ce6b8f31-afee-4062-b860-d054c4d37be0" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 596.875838] env[62385]: DEBUG nova.network.neutron [req-68463ccc-a535-4918-8838-4d077262c343 req-61ab5d98-b5a0-4aad-81d4-e603f8ef7e5f service nova] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Refreshing network info cache for port e057457e-a781-4fd0-b8da-a1ea032ef872 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 596.961263] env[62385]: ERROR nova.compute.manager [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e057457e-a781-4fd0-b8da-a1ea032ef872, please check neutron logs for more information. [ 596.961263] env[62385]: ERROR nova.compute.manager Traceback (most recent call last): [ 596.961263] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 596.961263] env[62385]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 596.961263] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 596.961263] env[62385]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 596.961263] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 596.961263] env[62385]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 596.961263] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.961263] env[62385]: ERROR nova.compute.manager self.force_reraise() [ 596.961263] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.961263] env[62385]: ERROR nova.compute.manager raise self.value [ 596.961263] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 596.961263] env[62385]: ERROR nova.compute.manager updated_port = self._update_port( [ 596.961263] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.961263] env[62385]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 596.961623] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.961623] env[62385]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 596.961623] env[62385]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e057457e-a781-4fd0-b8da-a1ea032ef872, please check neutron logs for more information. [ 596.961623] env[62385]: ERROR nova.compute.manager [ 596.961623] env[62385]: Traceback (most recent call last): [ 596.961623] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 596.961623] env[62385]: listener.cb(fileno) [ 596.961623] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 596.961623] env[62385]: result = function(*args, **kwargs) [ 596.961623] env[62385]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 596.961623] env[62385]: return func(*args, **kwargs) [ 596.961623] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 596.961623] env[62385]: raise e [ 596.961623] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 596.961623] env[62385]: nwinfo = self.network_api.allocate_for_instance( [ 596.961623] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 596.961623] env[62385]: created_port_ids = self._update_ports_for_instance( [ 596.961623] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 596.961623] env[62385]: with excutils.save_and_reraise_exception(): [ 596.961623] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.961623] env[62385]: self.force_reraise() [ 596.961623] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.961623] env[62385]: raise self.value [ 596.961623] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 596.961623] env[62385]: updated_port = self._update_port( [ 596.961623] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.961623] env[62385]: _ensure_no_port_binding_failure(port) [ 596.961623] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.961623] env[62385]: raise exception.PortBindingFailed(port_id=port['id']) [ 596.962377] env[62385]: nova.exception.PortBindingFailed: Binding failed for port e057457e-a781-4fd0-b8da-a1ea032ef872, please check neutron logs for more information. [ 596.962377] env[62385]: Removing descriptor: 17 [ 597.155858] env[62385]: DEBUG nova.compute.manager [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 597.181649] env[62385]: DEBUG nova.virt.hardware [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 597.181892] env[62385]: DEBUG nova.virt.hardware [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 597.182057] env[62385]: DEBUG nova.virt.hardware [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 597.182242] env[62385]: DEBUG nova.virt.hardware [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 597.182390] env[62385]: DEBUG nova.virt.hardware [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 597.182533] env[62385]: DEBUG nova.virt.hardware [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 597.182736] env[62385]: DEBUG nova.virt.hardware [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 597.182891] env[62385]: DEBUG nova.virt.hardware [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 597.183188] env[62385]: DEBUG nova.virt.hardware [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 597.183426] env[62385]: DEBUG nova.virt.hardware [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 597.183662] env[62385]: DEBUG nova.virt.hardware [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 597.184643] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c7a7c83-ead2-4051-9b65-01632f0ebd8f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.192463] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa910161-49d4-4b6e-a48e-cb0b167991c8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.206501] env[62385]: ERROR nova.compute.manager [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e057457e-a781-4fd0-b8da-a1ea032ef872, please check neutron logs for more information. [ 597.206501] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Traceback (most recent call last): [ 597.206501] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 597.206501] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] yield resources [ 597.206501] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 597.206501] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] self.driver.spawn(context, instance, image_meta, [ 597.206501] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 597.206501] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 597.206501] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 597.206501] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] vm_ref = self.build_virtual_machine(instance, [ 597.206501] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 597.206797] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] vif_infos = vmwarevif.get_vif_info(self._session, [ 597.206797] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 597.206797] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] for vif in network_info: [ 597.206797] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 597.206797] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] return self._sync_wrapper(fn, *args, **kwargs) [ 597.206797] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 597.206797] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] self.wait() [ 597.206797] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 597.206797] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] self[:] = self._gt.wait() [ 597.206797] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 597.206797] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] return self._exit_event.wait() [ 597.206797] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 597.206797] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] current.throw(*self._exc) [ 597.207137] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 597.207137] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] result = function(*args, **kwargs) [ 597.207137] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 597.207137] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] return func(*args, **kwargs) [ 597.207137] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 597.207137] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] raise e [ 597.207137] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 597.207137] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] nwinfo = self.network_api.allocate_for_instance( [ 597.207137] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 597.207137] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] created_port_ids = self._update_ports_for_instance( [ 597.207137] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 597.207137] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] with excutils.save_and_reraise_exception(): [ 597.207137] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.207507] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] self.force_reraise() [ 597.207507] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.207507] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] raise self.value [ 597.207507] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 597.207507] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] updated_port = self._update_port( [ 597.207507] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.207507] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] _ensure_no_port_binding_failure(port) [ 597.207507] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.207507] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] raise exception.PortBindingFailed(port_id=port['id']) [ 597.207507] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] nova.exception.PortBindingFailed: Binding failed for port e057457e-a781-4fd0-b8da-a1ea032ef872, please check neutron logs for more information. [ 597.207507] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] [ 597.207507] env[62385]: INFO nova.compute.manager [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Terminating instance [ 597.209697] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "refresh_cache-ce6b8f31-afee-4062-b860-d054c4d37be0" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.312807] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.185s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.313626] env[62385]: ERROR nova.compute.manager [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 72a53ea0-2518-4f4e-8184-42d5b63b7147, please check neutron logs for more information. [ 597.313626] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Traceback (most recent call last): [ 597.313626] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 597.313626] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] self.driver.spawn(context, instance, image_meta, [ 597.313626] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 597.313626] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 597.313626] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 597.313626] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] vm_ref = self.build_virtual_machine(instance, [ 597.313626] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 597.313626] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] vif_infos = vmwarevif.get_vif_info(self._session, [ 597.313626] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 597.313897] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] for vif in network_info: [ 597.313897] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 597.313897] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] return self._sync_wrapper(fn, *args, **kwargs) [ 597.313897] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 597.313897] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] self.wait() [ 597.313897] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 597.313897] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] self[:] = self._gt.wait() [ 597.313897] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 597.313897] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] return self._exit_event.wait() [ 597.313897] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 597.313897] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] result = hub.switch() [ 597.313897] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 597.313897] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] return self.greenlet.switch() [ 597.314204] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 597.314204] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] result = function(*args, **kwargs) [ 597.314204] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 597.314204] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] return func(*args, **kwargs) [ 597.314204] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 597.314204] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] raise e [ 597.314204] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 597.314204] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] nwinfo = self.network_api.allocate_for_instance( [ 597.314204] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 597.314204] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] created_port_ids = self._update_ports_for_instance( [ 597.314204] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 597.314204] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] with excutils.save_and_reraise_exception(): [ 597.314204] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.314475] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] self.force_reraise() [ 597.314475] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.314475] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] raise self.value [ 597.314475] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 597.314475] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] updated_port = self._update_port( [ 597.314475] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.314475] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] _ensure_no_port_binding_failure(port) [ 597.314475] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.314475] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] raise exception.PortBindingFailed(port_id=port['id']) [ 597.314475] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] nova.exception.PortBindingFailed: Binding failed for port 72a53ea0-2518-4f4e-8184-42d5b63b7147, please check neutron logs for more information. [ 597.314475] env[62385]: ERROR nova.compute.manager [instance: 759d9575-84bf-4dec-846a-e63126dae49c] [ 597.314705] env[62385]: DEBUG nova.compute.utils [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Binding failed for port 72a53ea0-2518-4f4e-8184-42d5b63b7147, please check neutron logs for more information. {{(pid=62385) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 597.315996] env[62385]: DEBUG oslo_concurrency.lockutils [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.388s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.320197] env[62385]: DEBUG nova.compute.manager [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Build of instance 759d9575-84bf-4dec-846a-e63126dae49c was re-scheduled: Binding failed for port 72a53ea0-2518-4f4e-8184-42d5b63b7147, please check neutron logs for more information. {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 597.320687] env[62385]: DEBUG nova.compute.manager [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Unplugging VIFs for instance {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 597.321043] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] Acquiring lock "refresh_cache-759d9575-84bf-4dec-846a-e63126dae49c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.321301] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] Acquired lock "refresh_cache-759d9575-84bf-4dec-846a-e63126dae49c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.321521] env[62385]: DEBUG nova.network.neutron [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 597.353116] env[62385]: DEBUG oslo_concurrency.lockutils [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.397456] env[62385]: DEBUG nova.network.neutron [req-68463ccc-a535-4918-8838-4d077262c343 req-61ab5d98-b5a0-4aad-81d4-e603f8ef7e5f service nova] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 597.427866] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 597.427866] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 597.484078] env[62385]: DEBUG nova.network.neutron [req-68463ccc-a535-4918-8838-4d077262c343 req-61ab5d98-b5a0-4aad-81d4-e603f8ef7e5f service nova] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.853122] env[62385]: DEBUG nova.network.neutron [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 597.942938] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 597.943148] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Starting heal instance info cache {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 597.943277] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Rebuilding the list of instances to heal {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 597.987650] env[62385]: DEBUG oslo_concurrency.lockutils [req-68463ccc-a535-4918-8838-4d077262c343 req-61ab5d98-b5a0-4aad-81d4-e603f8ef7e5f service nova] Releasing lock "refresh_cache-ce6b8f31-afee-4062-b860-d054c4d37be0" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 597.988107] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquired lock "refresh_cache-ce6b8f31-afee-4062-b860-d054c4d37be0" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.988294] env[62385]: DEBUG nova.network.neutron [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 598.021028] env[62385]: DEBUG nova.network.neutron [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.348448] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60866788-1f6f-4672-8e33-57372f199bc4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.356572] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d4d199e-8c7e-47fb-ba39-f8dd64160430 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.390557] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a356672d-f664-4362-b09f-248945f024b2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.398634] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cae87b1-f1aa-492b-aaf6-ed0c64a15fe8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.412405] env[62385]: DEBUG nova.compute.provider_tree [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 598.452026] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Skipping network cache update for instance because it is Building. {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 598.452026] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Skipping network cache update for instance because it is Building. {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 598.452026] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Skipping network cache update for instance because it is Building. {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 598.487797] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquiring lock "refresh_cache-e514a72d-6b4d-46f9-935b-3535bed4afc2" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 598.488581] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquired lock "refresh_cache-e514a72d-6b4d-46f9-935b-3535bed4afc2" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.488581] env[62385]: DEBUG nova.network.neutron [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Forcefully refreshing network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 598.488581] env[62385]: DEBUG nova.objects.instance [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lazy-loading 'info_cache' on Instance uuid e514a72d-6b4d-46f9-935b-3535bed4afc2 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 598.514444] env[62385]: DEBUG nova.network.neutron [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 598.529377] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] Releasing lock "refresh_cache-759d9575-84bf-4dec-846a-e63126dae49c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 598.529377] env[62385]: DEBUG nova.compute.manager [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 598.529998] env[62385]: DEBUG nova.compute.manager [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 598.529998] env[62385]: DEBUG nova.network.neutron [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 598.557246] env[62385]: DEBUG nova.network.neutron [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 598.640190] env[62385]: DEBUG nova.network.neutron [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.899367] env[62385]: DEBUG nova.compute.manager [req-653eb792-9feb-49e6-8544-3a9fb991485e req-012a12d4-44c1-4de3-be24-a086f61a23cd service nova] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Received event network-vif-deleted-e057457e-a781-4fd0-b8da-a1ea032ef872 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 598.916384] env[62385]: DEBUG nova.scheduler.client.report [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 599.059445] env[62385]: DEBUG nova.network.neutron [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.143552] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Releasing lock "refresh_cache-ce6b8f31-afee-4062-b860-d054c4d37be0" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.144140] env[62385]: DEBUG nova.compute.manager [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 599.144338] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 599.144636] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-adc82253-1606-4f50-a7cf-2b73a7cf9bb7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.154446] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ea0c6ce-be22-445e-a5bd-8cfef7c978f0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.179180] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ce6b8f31-afee-4062-b860-d054c4d37be0 could not be found. [ 599.179403] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 599.179584] env[62385]: INFO nova.compute.manager [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Took 0.04 seconds to destroy the instance on the hypervisor. [ 599.179819] env[62385]: DEBUG oslo.service.loopingcall [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 599.180040] env[62385]: DEBUG nova.compute.manager [-] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 599.180137] env[62385]: DEBUG nova.network.neutron [-] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 599.194966] env[62385]: DEBUG nova.network.neutron [-] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 599.421681] env[62385]: DEBUG oslo_concurrency.lockutils [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.106s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 599.422818] env[62385]: ERROR nova.compute.manager [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2dddce72-7166-457f-8a3a-f4139ac7cca4, please check neutron logs for more information. [ 599.422818] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Traceback (most recent call last): [ 599.422818] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 599.422818] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] self.driver.spawn(context, instance, image_meta, [ 599.422818] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 599.422818] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] self._vmops.spawn(context, instance, image_meta, injected_files, [ 599.422818] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 599.422818] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] vm_ref = self.build_virtual_machine(instance, [ 599.422818] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 599.422818] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] vif_infos = vmwarevif.get_vif_info(self._session, [ 599.422818] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 599.424386] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] for vif in network_info: [ 599.424386] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 599.424386] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] return self._sync_wrapper(fn, *args, **kwargs) [ 599.424386] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 599.424386] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] self.wait() [ 599.424386] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 599.424386] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] self[:] = self._gt.wait() [ 599.424386] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 599.424386] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] return self._exit_event.wait() [ 599.424386] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 599.424386] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] result = hub.switch() [ 599.424386] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 599.424386] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] return self.greenlet.switch() [ 599.424944] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 599.424944] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] result = function(*args, **kwargs) [ 599.424944] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 599.424944] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] return func(*args, **kwargs) [ 599.424944] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 599.424944] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] raise e [ 599.424944] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 599.424944] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] nwinfo = self.network_api.allocate_for_instance( [ 599.424944] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 599.424944] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] created_port_ids = self._update_ports_for_instance( [ 599.424944] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 599.424944] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] with excutils.save_and_reraise_exception(): [ 599.424944] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.425280] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] self.force_reraise() [ 599.425280] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.425280] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] raise self.value [ 599.425280] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 599.425280] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] updated_port = self._update_port( [ 599.425280] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.425280] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] _ensure_no_port_binding_failure(port) [ 599.425280] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.425280] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] raise exception.PortBindingFailed(port_id=port['id']) [ 599.425280] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] nova.exception.PortBindingFailed: Binding failed for port 2dddce72-7166-457f-8a3a-f4139ac7cca4, please check neutron logs for more information. [ 599.425280] env[62385]: ERROR nova.compute.manager [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] [ 599.425524] env[62385]: DEBUG nova.compute.utils [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Binding failed for port 2dddce72-7166-457f-8a3a-f4139ac7cca4, please check neutron logs for more information. {{(pid=62385) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 599.425524] env[62385]: DEBUG oslo_concurrency.lockutils [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.801s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.426452] env[62385]: INFO nova.compute.claims [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 599.431179] env[62385]: DEBUG nova.compute.manager [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Build of instance be595c9e-703e-4246-ae1c-2c77b240cb19 was re-scheduled: Binding failed for port 2dddce72-7166-457f-8a3a-f4139ac7cca4, please check neutron logs for more information. {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 599.432586] env[62385]: DEBUG nova.compute.manager [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Unplugging VIFs for instance {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 599.432844] env[62385]: DEBUG oslo_concurrency.lockutils [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] Acquiring lock "refresh_cache-be595c9e-703e-4246-ae1c-2c77b240cb19" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 599.432989] env[62385]: DEBUG oslo_concurrency.lockutils [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] Acquired lock "refresh_cache-be595c9e-703e-4246-ae1c-2c77b240cb19" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 599.433190] env[62385]: DEBUG nova.network.neutron [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 599.523832] env[62385]: DEBUG nova.network.neutron [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 599.562886] env[62385]: INFO nova.compute.manager [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] [instance: 759d9575-84bf-4dec-846a-e63126dae49c] Took 1.03 seconds to deallocate network for instance. [ 599.697697] env[62385]: DEBUG nova.network.neutron [-] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.957222] env[62385]: DEBUG nova.network.neutron [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 600.042495] env[62385]: DEBUG nova.network.neutron [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.174293] env[62385]: DEBUG nova.network.neutron [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.201598] env[62385]: INFO nova.compute.manager [-] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Took 1.02 seconds to deallocate network for instance. [ 600.206478] env[62385]: DEBUG nova.compute.claims [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Aborting claim: {{(pid=62385) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 600.206478] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.547323] env[62385]: DEBUG oslo_concurrency.lockutils [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] Releasing lock "refresh_cache-be595c9e-703e-4246-ae1c-2c77b240cb19" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 600.547323] env[62385]: DEBUG nova.compute.manager [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 600.547323] env[62385]: DEBUG nova.compute.manager [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 600.547323] env[62385]: DEBUG nova.network.neutron [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 600.587282] env[62385]: DEBUG nova.network.neutron [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 600.605822] env[62385]: INFO nova.scheduler.client.report [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] Deleted allocations for instance 759d9575-84bf-4dec-846a-e63126dae49c [ 600.679672] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Releasing lock "refresh_cache-e514a72d-6b4d-46f9-935b-3535bed4afc2" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 600.679883] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Updated the network info_cache for instance {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 600.680105] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 600.680268] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 600.680444] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 600.680621] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 600.680763] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 600.680907] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 600.681052] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62385) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 600.681775] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager.update_available_resource {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 600.858158] env[62385]: DEBUG nova.compute.manager [None req-8ac56b00-37fa-46c8-9d21-d6e52a23ef2b tempest-ServerDiagnosticsV248Test-929408931 tempest-ServerDiagnosticsV248Test-929408931-project-admin] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 600.859862] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab5d8fa9-aee9-4724-8a59-82a42be79018 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.866839] env[62385]: INFO nova.compute.manager [None req-8ac56b00-37fa-46c8-9d21-d6e52a23ef2b tempest-ServerDiagnosticsV248Test-929408931 tempest-ServerDiagnosticsV248Test-929408931-project-admin] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Retrieving diagnostics [ 600.867835] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc0e6fc0-3c65-4683-9d0c-a499b022824b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.092488] env[62385]: DEBUG nova.network.neutron [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.117208] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2cf24874-9899-4293-8864-73e4391917b6 tempest-ServersTestJSON-1364563416 tempest-ServersTestJSON-1364563416-project-member] Lock "759d9575-84bf-4dec-846a-e63126dae49c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 68.436s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 601.149012] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cac19d71-890b-49d6-bbc9-264771602794 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.156990] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a89dad29-c602-4f47-a957-352a4fac2178 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.189942] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 601.189942] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6068d43a-c9ac-49b0-ab75-fefa897791b5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.196588] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b38cc042-9c6f-47aa-a48b-be1880ff17a7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.211034] env[62385]: DEBUG nova.compute.provider_tree [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 601.594533] env[62385]: INFO nova.compute.manager [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] [instance: be595c9e-703e-4246-ae1c-2c77b240cb19] Took 1.05 seconds to deallocate network for instance. [ 601.623565] env[62385]: DEBUG nova.compute.manager [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 601.715603] env[62385]: DEBUG nova.scheduler.client.report [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 602.152100] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.223329] env[62385]: DEBUG oslo_concurrency.lockutils [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.798s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.223849] env[62385]: DEBUG nova.compute.manager [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 602.226342] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.415s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.273679] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3d83202f-9f92-4596-a521-3507c75c3596 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Acquiring lock "e514a72d-6b4d-46f9-935b-3535bed4afc2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.273936] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3d83202f-9f92-4596-a521-3507c75c3596 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Lock "e514a72d-6b4d-46f9-935b-3535bed4afc2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.274270] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3d83202f-9f92-4596-a521-3507c75c3596 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Acquiring lock "e514a72d-6b4d-46f9-935b-3535bed4afc2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.274363] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3d83202f-9f92-4596-a521-3507c75c3596 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Lock "e514a72d-6b4d-46f9-935b-3535bed4afc2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.274527] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3d83202f-9f92-4596-a521-3507c75c3596 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Lock "e514a72d-6b4d-46f9-935b-3535bed4afc2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.276544] env[62385]: INFO nova.compute.manager [None req-3d83202f-9f92-4596-a521-3507c75c3596 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Terminating instance [ 602.278050] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3d83202f-9f92-4596-a521-3507c75c3596 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Acquiring lock "refresh_cache-e514a72d-6b4d-46f9-935b-3535bed4afc2" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.278207] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3d83202f-9f92-4596-a521-3507c75c3596 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Acquired lock "refresh_cache-e514a72d-6b4d-46f9-935b-3535bed4afc2" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.278374] env[62385]: DEBUG nova.network.neutron [None req-3d83202f-9f92-4596-a521-3507c75c3596 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 602.644015] env[62385]: INFO nova.scheduler.client.report [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] Deleted allocations for instance be595c9e-703e-4246-ae1c-2c77b240cb19 [ 602.732264] env[62385]: DEBUG nova.compute.utils [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 602.737619] env[62385]: DEBUG nova.compute.manager [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 602.739268] env[62385]: DEBUG nova.network.neutron [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 602.803189] env[62385]: DEBUG nova.network.neutron [None req-3d83202f-9f92-4596-a521-3507c75c3596 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 602.819154] env[62385]: DEBUG nova.policy [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dfccdfd301334545aa0eaa796a955c43', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2ebc1c1ed0bb404b89e59325535edecd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 602.890954] env[62385]: DEBUG nova.network.neutron [None req-3d83202f-9f92-4596-a521-3507c75c3596 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.159839] env[62385]: DEBUG oslo_concurrency.lockutils [None req-dd52b59b-a7bb-4867-9ccb-a893ca65bfa6 tempest-ServersWithSpecificFlavorTestJSON-260805524 tempest-ServersWithSpecificFlavorTestJSON-260805524-project-member] Lock "be595c9e-703e-4246-ae1c-2c77b240cb19" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 68.213s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 603.229016] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Acquiring lock "92057af7-28a3-4643-9cda-d3d868d0cfc9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.229252] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Lock "92057af7-28a3-4643-9cda-d3d868d0cfc9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.234961] env[62385]: DEBUG nova.network.neutron [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Successfully created port: c7bdc405-0f30-4e93-a8e5-daffaabdfaf8 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 603.238921] env[62385]: DEBUG nova.compute.manager [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 603.260356] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44aca1b7-6e53-4b73-8bd9-8401ae32e4e8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.873091] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3d83202f-9f92-4596-a521-3507c75c3596 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Releasing lock "refresh_cache-e514a72d-6b4d-46f9-935b-3535bed4afc2" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.873830] env[62385]: DEBUG nova.compute.manager [None req-3d83202f-9f92-4596-a521-3507c75c3596 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 603.873830] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3d83202f-9f92-4596-a521-3507c75c3596 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 603.874280] env[62385]: DEBUG nova.compute.manager [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 603.882942] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52de092a-5807-4715-94b0-7bdc0e13c0c1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.883817] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a3303b3-e8f0-493c-99ee-56025ca59450 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.892763] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d83202f-9f92-4596-a521-3507c75c3596 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 603.921217] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e583a91b-375e-40ed-b4c8-187f05703f79 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.924711] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2bf0d15-b7cb-48c3-9906-0523174ac87d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.933868] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b61dd08-97d4-48dc-ad75-15a670b11c31 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.940030] env[62385]: DEBUG oslo_vmware.api [None req-3d83202f-9f92-4596-a521-3507c75c3596 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Waiting for the task: (returnval){ [ 603.940030] env[62385]: value = "task-1205445" [ 603.940030] env[62385]: _type = "Task" [ 603.940030] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.952613] env[62385]: DEBUG nova.compute.provider_tree [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 603.958625] env[62385]: DEBUG oslo_vmware.api [None req-3d83202f-9f92-4596-a521-3507c75c3596 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Task: {'id': task-1205445, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.390789] env[62385]: DEBUG nova.compute.manager [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 604.402225] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.420080] env[62385]: DEBUG nova.virt.hardware [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 604.420330] env[62385]: DEBUG nova.virt.hardware [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 604.420576] env[62385]: DEBUG nova.virt.hardware [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 604.420769] env[62385]: DEBUG nova.virt.hardware [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 604.420911] env[62385]: DEBUG nova.virt.hardware [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 604.421457] env[62385]: DEBUG nova.virt.hardware [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 604.421712] env[62385]: DEBUG nova.virt.hardware [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 604.421879] env[62385]: DEBUG nova.virt.hardware [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 604.422092] env[62385]: DEBUG nova.virt.hardware [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 604.422314] env[62385]: DEBUG nova.virt.hardware [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 604.422503] env[62385]: DEBUG nova.virt.hardware [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 604.423644] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8322fbb-ea4f-4ed3-a197-13c84a7229f3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.433386] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b91dd73-0b05-4fc3-b92a-3569a051942d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.456671] env[62385]: DEBUG nova.scheduler.client.report [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 604.459937] env[62385]: DEBUG oslo_vmware.api [None req-3d83202f-9f92-4596-a521-3507c75c3596 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Task: {'id': task-1205445, 'name': PowerOffVM_Task, 'duration_secs': 0.13484} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 604.460399] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d83202f-9f92-4596-a521-3507c75c3596 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 604.460571] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3d83202f-9f92-4596-a521-3507c75c3596 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 604.464025] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d3248511-dd3a-4d38-ba7b-4c738bbb4d53 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.486969] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3d83202f-9f92-4596-a521-3507c75c3596 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 604.487238] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3d83202f-9f92-4596-a521-3507c75c3596 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 604.487447] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d83202f-9f92-4596-a521-3507c75c3596 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Deleting the datastore file [datastore1] e514a72d-6b4d-46f9-935b-3535bed4afc2 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 604.487685] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-38a1bc09-11c4-4ad0-915c-d3a9eff043c4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.495108] env[62385]: DEBUG oslo_vmware.api [None req-3d83202f-9f92-4596-a521-3507c75c3596 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Waiting for the task: (returnval){ [ 604.495108] env[62385]: value = "task-1205447" [ 604.495108] env[62385]: _type = "Task" [ 604.495108] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 604.502717] env[62385]: DEBUG oslo_vmware.api [None req-3d83202f-9f92-4596-a521-3507c75c3596 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Task: {'id': task-1205447, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.831153] env[62385]: DEBUG nova.compute.manager [req-a19945e7-6359-4de9-9a58-4e0ea1b90792 req-bc37cb2f-d837-4a11-9c7b-8b7ee86505d0 service nova] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Received event network-changed-c7bdc405-0f30-4e93-a8e5-daffaabdfaf8 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 604.831153] env[62385]: DEBUG nova.compute.manager [req-a19945e7-6359-4de9-9a58-4e0ea1b90792 req-bc37cb2f-d837-4a11-9c7b-8b7ee86505d0 service nova] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Refreshing instance network info cache due to event network-changed-c7bdc405-0f30-4e93-a8e5-daffaabdfaf8. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 604.831153] env[62385]: DEBUG oslo_concurrency.lockutils [req-a19945e7-6359-4de9-9a58-4e0ea1b90792 req-bc37cb2f-d837-4a11-9c7b-8b7ee86505d0 service nova] Acquiring lock "refresh_cache-4a99d2e8-7795-4016-ad6e-39a816c7fae1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.831153] env[62385]: DEBUG oslo_concurrency.lockutils [req-a19945e7-6359-4de9-9a58-4e0ea1b90792 req-bc37cb2f-d837-4a11-9c7b-8b7ee86505d0 service nova] Acquired lock "refresh_cache-4a99d2e8-7795-4016-ad6e-39a816c7fae1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.831153] env[62385]: DEBUG nova.network.neutron [req-a19945e7-6359-4de9-9a58-4e0ea1b90792 req-bc37cb2f-d837-4a11-9c7b-8b7ee86505d0 service nova] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Refreshing network info cache for port c7bdc405-0f30-4e93-a8e5-daffaabdfaf8 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 604.961420] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.735s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 604.962065] env[62385]: ERROR nova.compute.manager [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f46e1314-833f-428a-960e-f4c5bbc35461, please check neutron logs for more information. [ 604.962065] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Traceback (most recent call last): [ 604.962065] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 604.962065] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] self.driver.spawn(context, instance, image_meta, [ 604.962065] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 604.962065] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] self._vmops.spawn(context, instance, image_meta, injected_files, [ 604.962065] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 604.962065] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] vm_ref = self.build_virtual_machine(instance, [ 604.962065] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 604.962065] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] vif_infos = vmwarevif.get_vif_info(self._session, [ 604.962065] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 604.962390] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] for vif in network_info: [ 604.962390] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 604.962390] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] return self._sync_wrapper(fn, *args, **kwargs) [ 604.962390] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 604.962390] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] self.wait() [ 604.962390] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 604.962390] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] self[:] = self._gt.wait() [ 604.962390] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 604.962390] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] return self._exit_event.wait() [ 604.962390] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 604.962390] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] current.throw(*self._exc) [ 604.962390] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 604.962390] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] result = function(*args, **kwargs) [ 604.962737] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 604.962737] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] return func(*args, **kwargs) [ 604.962737] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 604.962737] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] raise e [ 604.962737] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.962737] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] nwinfo = self.network_api.allocate_for_instance( [ 604.962737] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 604.962737] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] created_port_ids = self._update_ports_for_instance( [ 604.962737] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 604.962737] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] with excutils.save_and_reraise_exception(): [ 604.962737] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.962737] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] self.force_reraise() [ 604.962737] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.963099] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] raise self.value [ 604.963099] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 604.963099] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] updated_port = self._update_port( [ 604.963099] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.963099] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] _ensure_no_port_binding_failure(port) [ 604.963099] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.963099] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] raise exception.PortBindingFailed(port_id=port['id']) [ 604.963099] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] nova.exception.PortBindingFailed: Binding failed for port f46e1314-833f-428a-960e-f4c5bbc35461, please check neutron logs for more information. [ 604.963099] env[62385]: ERROR nova.compute.manager [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] [ 604.963099] env[62385]: DEBUG nova.compute.utils [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Binding failed for port f46e1314-833f-428a-960e-f4c5bbc35461, please check neutron logs for more information. {{(pid=62385) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 604.964145] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.627s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.967312] env[62385]: INFO nova.compute.claims [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 604.971646] env[62385]: DEBUG nova.compute.manager [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Build of instance 0f1d83e7-bb40-4344-a344-2fc98efe8528 was re-scheduled: Binding failed for port f46e1314-833f-428a-960e-f4c5bbc35461, please check neutron logs for more information. {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 604.972323] env[62385]: DEBUG nova.compute.manager [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Unplugging VIFs for instance {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 604.972802] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] Acquiring lock "refresh_cache-0f1d83e7-bb40-4344-a344-2fc98efe8528" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.973035] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] Acquired lock "refresh_cache-0f1d83e7-bb40-4344-a344-2fc98efe8528" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.973281] env[62385]: DEBUG nova.network.neutron [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 605.004587] env[62385]: DEBUG oslo_vmware.api [None req-3d83202f-9f92-4596-a521-3507c75c3596 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Task: {'id': task-1205447, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.097285} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 605.004935] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d83202f-9f92-4596-a521-3507c75c3596 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 605.005306] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3d83202f-9f92-4596-a521-3507c75c3596 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 605.005534] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3d83202f-9f92-4596-a521-3507c75c3596 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 605.006287] env[62385]: INFO nova.compute.manager [None req-3d83202f-9f92-4596-a521-3507c75c3596 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Took 1.13 seconds to destroy the instance on the hypervisor. [ 605.006658] env[62385]: DEBUG oslo.service.loopingcall [None req-3d83202f-9f92-4596-a521-3507c75c3596 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 605.006862] env[62385]: DEBUG nova.compute.manager [-] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 605.006958] env[62385]: DEBUG nova.network.neutron [-] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 605.025374] env[62385]: DEBUG nova.network.neutron [-] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 605.311599] env[62385]: ERROR nova.compute.manager [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c7bdc405-0f30-4e93-a8e5-daffaabdfaf8, please check neutron logs for more information. [ 605.311599] env[62385]: ERROR nova.compute.manager Traceback (most recent call last): [ 605.311599] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 605.311599] env[62385]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 605.311599] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 605.311599] env[62385]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 605.311599] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 605.311599] env[62385]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 605.311599] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.311599] env[62385]: ERROR nova.compute.manager self.force_reraise() [ 605.311599] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.311599] env[62385]: ERROR nova.compute.manager raise self.value [ 605.311599] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 605.311599] env[62385]: ERROR nova.compute.manager updated_port = self._update_port( [ 605.311599] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.311599] env[62385]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 605.312213] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.312213] env[62385]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 605.312213] env[62385]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c7bdc405-0f30-4e93-a8e5-daffaabdfaf8, please check neutron logs for more information. [ 605.312213] env[62385]: ERROR nova.compute.manager [ 605.312213] env[62385]: Traceback (most recent call last): [ 605.312213] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 605.312213] env[62385]: listener.cb(fileno) [ 605.312213] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 605.312213] env[62385]: result = function(*args, **kwargs) [ 605.312213] env[62385]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 605.312213] env[62385]: return func(*args, **kwargs) [ 605.312213] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 605.312213] env[62385]: raise e [ 605.312213] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 605.312213] env[62385]: nwinfo = self.network_api.allocate_for_instance( [ 605.312213] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 605.312213] env[62385]: created_port_ids = self._update_ports_for_instance( [ 605.312213] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 605.312213] env[62385]: with excutils.save_and_reraise_exception(): [ 605.312213] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.312213] env[62385]: self.force_reraise() [ 605.312213] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.312213] env[62385]: raise self.value [ 605.312213] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 605.312213] env[62385]: updated_port = self._update_port( [ 605.312213] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.312213] env[62385]: _ensure_no_port_binding_failure(port) [ 605.312213] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.312213] env[62385]: raise exception.PortBindingFailed(port_id=port['id']) [ 605.312885] env[62385]: nova.exception.PortBindingFailed: Binding failed for port c7bdc405-0f30-4e93-a8e5-daffaabdfaf8, please check neutron logs for more information. [ 605.312885] env[62385]: Removing descriptor: 19 [ 605.312885] env[62385]: ERROR nova.compute.manager [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c7bdc405-0f30-4e93-a8e5-daffaabdfaf8, please check neutron logs for more information. [ 605.312885] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Traceback (most recent call last): [ 605.312885] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 605.312885] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] yield resources [ 605.312885] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 605.312885] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] self.driver.spawn(context, instance, image_meta, [ 605.312885] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 605.312885] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 605.312885] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 605.312885] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] vm_ref = self.build_virtual_machine(instance, [ 605.313227] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 605.313227] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] vif_infos = vmwarevif.get_vif_info(self._session, [ 605.313227] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 605.313227] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] for vif in network_info: [ 605.313227] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 605.313227] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] return self._sync_wrapper(fn, *args, **kwargs) [ 605.313227] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 605.313227] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] self.wait() [ 605.313227] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 605.313227] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] self[:] = self._gt.wait() [ 605.313227] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 605.313227] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] return self._exit_event.wait() [ 605.313227] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 605.313534] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] result = hub.switch() [ 605.313534] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 605.313534] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] return self.greenlet.switch() [ 605.313534] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 605.313534] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] result = function(*args, **kwargs) [ 605.313534] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 605.313534] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] return func(*args, **kwargs) [ 605.313534] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 605.313534] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] raise e [ 605.313534] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 605.313534] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] nwinfo = self.network_api.allocate_for_instance( [ 605.313534] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 605.313534] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] created_port_ids = self._update_ports_for_instance( [ 605.313817] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 605.313817] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] with excutils.save_and_reraise_exception(): [ 605.313817] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.313817] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] self.force_reraise() [ 605.313817] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.313817] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] raise self.value [ 605.313817] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 605.313817] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] updated_port = self._update_port( [ 605.313817] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.313817] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] _ensure_no_port_binding_failure(port) [ 605.313817] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.313817] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] raise exception.PortBindingFailed(port_id=port['id']) [ 605.314092] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] nova.exception.PortBindingFailed: Binding failed for port c7bdc405-0f30-4e93-a8e5-daffaabdfaf8, please check neutron logs for more information. [ 605.314092] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] [ 605.314092] env[62385]: INFO nova.compute.manager [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Terminating instance [ 605.315424] env[62385]: DEBUG oslo_concurrency.lockutils [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Acquiring lock "refresh_cache-4a99d2e8-7795-4016-ad6e-39a816c7fae1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 605.367580] env[62385]: DEBUG nova.network.neutron [req-a19945e7-6359-4de9-9a58-4e0ea1b90792 req-bc37cb2f-d837-4a11-9c7b-8b7ee86505d0 service nova] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 605.501313] env[62385]: DEBUG nova.network.neutron [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 605.531400] env[62385]: DEBUG nova.network.neutron [-] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.541210] env[62385]: DEBUG nova.network.neutron [req-a19945e7-6359-4de9-9a58-4e0ea1b90792 req-bc37cb2f-d837-4a11-9c7b-8b7ee86505d0 service nova] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.595226] env[62385]: DEBUG nova.network.neutron [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.034517] env[62385]: INFO nova.compute.manager [-] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Took 1.03 seconds to deallocate network for instance. [ 606.051548] env[62385]: DEBUG oslo_concurrency.lockutils [req-a19945e7-6359-4de9-9a58-4e0ea1b90792 req-bc37cb2f-d837-4a11-9c7b-8b7ee86505d0 service nova] Releasing lock "refresh_cache-4a99d2e8-7795-4016-ad6e-39a816c7fae1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.051651] env[62385]: DEBUG oslo_concurrency.lockutils [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Acquired lock "refresh_cache-4a99d2e8-7795-4016-ad6e-39a816c7fae1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.051815] env[62385]: DEBUG nova.network.neutron [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 606.098076] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] Releasing lock "refresh_cache-0f1d83e7-bb40-4344-a344-2fc98efe8528" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.098076] env[62385]: DEBUG nova.compute.manager [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 606.098076] env[62385]: DEBUG nova.compute.manager [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 606.098252] env[62385]: DEBUG nova.network.neutron [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 606.117873] env[62385]: DEBUG nova.network.neutron [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 606.441720] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcdecea6-278f-4f93-a686-60b47242d4fd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.450021] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7902f720-e53c-4a0c-b574-27ff3bc4d982 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.481940] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fb2e1f0-319d-48c2-b1eb-844606ded926 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.489444] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e51bea8f-ff22-4330-834a-7111b1d73ffa {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.503908] env[62385]: DEBUG nova.compute.provider_tree [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 606.544565] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3d83202f-9f92-4596-a521-3507c75c3596 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.572337] env[62385]: DEBUG nova.network.neutron [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 606.620202] env[62385]: DEBUG nova.network.neutron [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.723093] env[62385]: DEBUG nova.network.neutron [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.999264] env[62385]: DEBUG nova.compute.manager [req-2b509a1c-33c2-4cf0-bf0d-fbddd4113d70 req-128acd47-04ae-4b2a-aa22-48569f6fec0c service nova] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Received event network-vif-deleted-c7bdc405-0f30-4e93-a8e5-daffaabdfaf8 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 607.004663] env[62385]: DEBUG nova.scheduler.client.report [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 607.122624] env[62385]: INFO nova.compute.manager [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] [instance: 0f1d83e7-bb40-4344-a344-2fc98efe8528] Took 1.02 seconds to deallocate network for instance. [ 607.226052] env[62385]: DEBUG oslo_concurrency.lockutils [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Releasing lock "refresh_cache-4a99d2e8-7795-4016-ad6e-39a816c7fae1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 607.226511] env[62385]: DEBUG nova.compute.manager [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 607.226703] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 607.227048] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-af31da71-127b-4bfe-ba29-39b388752ca5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.236961] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2713843-6ecf-4add-8741-1a0faee5bb4f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.260159] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4a99d2e8-7795-4016-ad6e-39a816c7fae1 could not be found. [ 607.260419] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 607.260599] env[62385]: INFO nova.compute.manager [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Took 0.03 seconds to destroy the instance on the hypervisor. [ 607.260834] env[62385]: DEBUG oslo.service.loopingcall [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 607.261077] env[62385]: DEBUG nova.compute.manager [-] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 607.261168] env[62385]: DEBUG nova.network.neutron [-] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 607.281416] env[62385]: DEBUG nova.network.neutron [-] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 607.510068] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.546s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 607.510399] env[62385]: DEBUG nova.compute.manager [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 607.513823] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.172s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 607.515257] env[62385]: INFO nova.compute.claims [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 607.784495] env[62385]: DEBUG nova.network.neutron [-] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.020703] env[62385]: DEBUG nova.compute.utils [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 608.023852] env[62385]: DEBUG nova.compute.manager [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 608.024057] env[62385]: DEBUG nova.network.neutron [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 608.069803] env[62385]: DEBUG nova.policy [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f9bbbfdd70ed419b8d65f027e59db0c2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a1524cc8cfd444e98436b96763444c8e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 608.166918] env[62385]: INFO nova.scheduler.client.report [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] Deleted allocations for instance 0f1d83e7-bb40-4344-a344-2fc98efe8528 [ 608.288367] env[62385]: INFO nova.compute.manager [-] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Took 1.03 seconds to deallocate network for instance. [ 608.291958] env[62385]: DEBUG nova.compute.claims [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Aborting claim: {{(pid=62385) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 608.292271] env[62385]: DEBUG oslo_concurrency.lockutils [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.422845] env[62385]: DEBUG nova.network.neutron [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Successfully created port: 1fac3d82-702d-4ce0-906e-0e251b78b679 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 608.524838] env[62385]: DEBUG nova.compute.manager [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 608.678167] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0306680c-0968-4016-81bf-f9631e7cfb5d tempest-ServerAddressesNegativeTestJSON-1282016917 tempest-ServerAddressesNegativeTestJSON-1282016917-project-member] Lock "0f1d83e7-bb40-4344-a344-2fc98efe8528" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.407s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.034086] env[62385]: INFO nova.virt.block_device [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Booting with volume 98319c2e-aa8a-40a2-8fd5-b1bd600c85f1 at /dev/sda [ 609.036817] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f3c106d-7268-46f6-b575-84bad503ea2f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.060734] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-258e6e8d-0e28-44db-b9c8-e8a358bcfce0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.101796] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c4fb35d-b2f0-4e8c-838a-b4d1e8b71bf4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.106120] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f4ee97ac-32da-4754-a49d-f5e4b43c7fed {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.116405] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b761b78f-c25c-488f-8e73-dd818486dcc3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.127280] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6cb9722-d212-499f-ab1b-ccec1782cb39 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.151294] env[62385]: DEBUG nova.compute.provider_tree [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 609.156302] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d30c83b8-7517-477c-932a-714dfde37d3e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.164538] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75a06714-3420-4a54-bb73-a63ba50e3288 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.181420] env[62385]: DEBUG nova.compute.manager [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 609.188739] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-309338a3-bdd4-4ce8-b88f-db77e37ce385 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.195106] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab8c309b-8525-41f3-8a71-3c82e3818e6b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.208785] env[62385]: DEBUG nova.virt.block_device [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Updating existing volume attachment record: e5bbfa92-97f6-49fb-845d-c6302bacbdc8 {{(pid=62385) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 609.403118] env[62385]: DEBUG nova.compute.manager [req-d931bc37-3fe1-43ff-981a-98f71a9518a0 req-7fa8c3c0-0d72-47f2-9f8a-c77b86ee365b service nova] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Received event network-changed-1fac3d82-702d-4ce0-906e-0e251b78b679 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 609.403118] env[62385]: DEBUG nova.compute.manager [req-d931bc37-3fe1-43ff-981a-98f71a9518a0 req-7fa8c3c0-0d72-47f2-9f8a-c77b86ee365b service nova] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Refreshing instance network info cache due to event network-changed-1fac3d82-702d-4ce0-906e-0e251b78b679. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 609.403118] env[62385]: DEBUG oslo_concurrency.lockutils [req-d931bc37-3fe1-43ff-981a-98f71a9518a0 req-7fa8c3c0-0d72-47f2-9f8a-c77b86ee365b service nova] Acquiring lock "refresh_cache-317e7e7c-7c6f-4855-a433-e55ea5571b92" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 609.404103] env[62385]: DEBUG oslo_concurrency.lockutils [req-d931bc37-3fe1-43ff-981a-98f71a9518a0 req-7fa8c3c0-0d72-47f2-9f8a-c77b86ee365b service nova] Acquired lock "refresh_cache-317e7e7c-7c6f-4855-a433-e55ea5571b92" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.404886] env[62385]: DEBUG nova.network.neutron [req-d931bc37-3fe1-43ff-981a-98f71a9518a0 req-7fa8c3c0-0d72-47f2-9f8a-c77b86ee365b service nova] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Refreshing network info cache for port 1fac3d82-702d-4ce0-906e-0e251b78b679 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 609.658046] env[62385]: DEBUG nova.scheduler.client.report [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 609.709010] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.851809] env[62385]: ERROR nova.compute.manager [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1fac3d82-702d-4ce0-906e-0e251b78b679, please check neutron logs for more information. [ 609.851809] env[62385]: ERROR nova.compute.manager Traceback (most recent call last): [ 609.851809] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.851809] env[62385]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 609.851809] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 609.851809] env[62385]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 609.851809] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 609.851809] env[62385]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 609.851809] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.851809] env[62385]: ERROR nova.compute.manager self.force_reraise() [ 609.851809] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.851809] env[62385]: ERROR nova.compute.manager raise self.value [ 609.851809] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 609.851809] env[62385]: ERROR nova.compute.manager updated_port = self._update_port( [ 609.851809] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.851809] env[62385]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 609.852468] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.852468] env[62385]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 609.852468] env[62385]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1fac3d82-702d-4ce0-906e-0e251b78b679, please check neutron logs for more information. [ 609.852468] env[62385]: ERROR nova.compute.manager [ 609.852468] env[62385]: Traceback (most recent call last): [ 609.852468] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 609.852468] env[62385]: listener.cb(fileno) [ 609.852468] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 609.852468] env[62385]: result = function(*args, **kwargs) [ 609.852468] env[62385]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 609.852468] env[62385]: return func(*args, **kwargs) [ 609.852468] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 609.852468] env[62385]: raise e [ 609.852468] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.852468] env[62385]: nwinfo = self.network_api.allocate_for_instance( [ 609.852468] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 609.852468] env[62385]: created_port_ids = self._update_ports_for_instance( [ 609.852468] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 609.852468] env[62385]: with excutils.save_and_reraise_exception(): [ 609.852468] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.852468] env[62385]: self.force_reraise() [ 609.852468] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.852468] env[62385]: raise self.value [ 609.852468] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 609.852468] env[62385]: updated_port = self._update_port( [ 609.852468] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.852468] env[62385]: _ensure_no_port_binding_failure(port) [ 609.852468] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.852468] env[62385]: raise exception.PortBindingFailed(port_id=port['id']) [ 609.855163] env[62385]: nova.exception.PortBindingFailed: Binding failed for port 1fac3d82-702d-4ce0-906e-0e251b78b679, please check neutron logs for more information. [ 609.855163] env[62385]: Removing descriptor: 17 [ 609.929293] env[62385]: DEBUG nova.network.neutron [req-d931bc37-3fe1-43ff-981a-98f71a9518a0 req-7fa8c3c0-0d72-47f2-9f8a-c77b86ee365b service nova] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 610.087926] env[62385]: DEBUG nova.network.neutron [req-d931bc37-3fe1-43ff-981a-98f71a9518a0 req-7fa8c3c0-0d72-47f2-9f8a-c77b86ee365b service nova] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.165803] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.652s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 610.166936] env[62385]: DEBUG nova.compute.manager [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 610.169592] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.475s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.591343] env[62385]: DEBUG oslo_concurrency.lockutils [req-d931bc37-3fe1-43ff-981a-98f71a9518a0 req-7fa8c3c0-0d72-47f2-9f8a-c77b86ee365b service nova] Releasing lock "refresh_cache-317e7e7c-7c6f-4855-a433-e55ea5571b92" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 610.672750] env[62385]: DEBUG nova.compute.utils [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 610.677705] env[62385]: DEBUG nova.compute.manager [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Not allocating networking since 'none' was specified. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 611.177949] env[62385]: DEBUG nova.compute.manager [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 611.181270] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98309688-6e3f-4b86-a694-15fc3f18303e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.189552] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93c6586a-db59-4262-9b89-9cfe29db4f29 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.222897] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a184917e-8f27-46e9-9040-f3565dd52fcd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.231485] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0348a526-9382-4f93-8e33-966eb29886a7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.246669] env[62385]: DEBUG nova.compute.provider_tree [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 611.330620] env[62385]: DEBUG nova.compute.manager [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 611.330620] env[62385]: DEBUG nova.virt.hardware [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 611.330620] env[62385]: DEBUG nova.virt.hardware [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 611.330620] env[62385]: DEBUG nova.virt.hardware [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 611.330872] env[62385]: DEBUG nova.virt.hardware [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 611.330872] env[62385]: DEBUG nova.virt.hardware [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 611.330872] env[62385]: DEBUG nova.virt.hardware [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 611.330872] env[62385]: DEBUG nova.virt.hardware [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 611.330872] env[62385]: DEBUG nova.virt.hardware [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 611.330998] env[62385]: DEBUG nova.virt.hardware [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 611.331253] env[62385]: DEBUG nova.virt.hardware [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 611.331438] env[62385]: DEBUG nova.virt.hardware [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 611.333095] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40877fe4-1ad9-45c6-bf6f-c118c4a7e417 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.340918] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73713f65-1eae-4a29-af83-dab6f5fdb9bd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.356101] env[62385]: ERROR nova.compute.manager [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1fac3d82-702d-4ce0-906e-0e251b78b679, please check neutron logs for more information. [ 611.356101] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Traceback (most recent call last): [ 611.356101] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 611.356101] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] yield resources [ 611.356101] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 611.356101] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] self.driver.spawn(context, instance, image_meta, [ 611.356101] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 611.356101] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] self._vmops.spawn(context, instance, image_meta, injected_files, [ 611.356101] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 611.356101] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] vm_ref = self.build_virtual_machine(instance, [ 611.356101] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 611.356698] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] vif_infos = vmwarevif.get_vif_info(self._session, [ 611.356698] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 611.356698] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] for vif in network_info: [ 611.356698] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 611.356698] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] return self._sync_wrapper(fn, *args, **kwargs) [ 611.356698] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 611.356698] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] self.wait() [ 611.356698] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 611.356698] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] self[:] = self._gt.wait() [ 611.356698] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 611.356698] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] return self._exit_event.wait() [ 611.356698] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 611.356698] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] current.throw(*self._exc) [ 611.357044] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.357044] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] result = function(*args, **kwargs) [ 611.357044] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 611.357044] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] return func(*args, **kwargs) [ 611.357044] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 611.357044] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] raise e [ 611.357044] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.357044] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] nwinfo = self.network_api.allocate_for_instance( [ 611.357044] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 611.357044] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] created_port_ids = self._update_ports_for_instance( [ 611.357044] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 611.357044] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] with excutils.save_and_reraise_exception(): [ 611.357044] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.357414] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] self.force_reraise() [ 611.357414] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.357414] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] raise self.value [ 611.357414] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 611.357414] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] updated_port = self._update_port( [ 611.357414] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.357414] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] _ensure_no_port_binding_failure(port) [ 611.357414] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.357414] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] raise exception.PortBindingFailed(port_id=port['id']) [ 611.357414] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] nova.exception.PortBindingFailed: Binding failed for port 1fac3d82-702d-4ce0-906e-0e251b78b679, please check neutron logs for more information. [ 611.357414] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] [ 611.357414] env[62385]: INFO nova.compute.manager [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Terminating instance [ 611.358499] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] Acquiring lock "refresh_cache-317e7e7c-7c6f-4855-a433-e55ea5571b92" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.359565] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] Acquired lock "refresh_cache-317e7e7c-7c6f-4855-a433-e55ea5571b92" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.359565] env[62385]: DEBUG nova.network.neutron [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 611.430582] env[62385]: DEBUG nova.compute.manager [req-db8a67ca-1ccc-4c0a-9184-b5bea118f418 req-0a48c1d2-73bd-4abc-9d53-9d444e686631 service nova] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Received event network-vif-deleted-1fac3d82-702d-4ce0-906e-0e251b78b679 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 611.750848] env[62385]: DEBUG nova.scheduler.client.report [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 611.877727] env[62385]: DEBUG nova.network.neutron [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 611.961870] env[62385]: DEBUG nova.network.neutron [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.189393] env[62385]: DEBUG nova.compute.manager [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 612.214306] env[62385]: DEBUG nova.virt.hardware [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 612.214552] env[62385]: DEBUG nova.virt.hardware [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 612.214713] env[62385]: DEBUG nova.virt.hardware [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 612.214892] env[62385]: DEBUG nova.virt.hardware [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 612.215048] env[62385]: DEBUG nova.virt.hardware [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 612.215202] env[62385]: DEBUG nova.virt.hardware [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 612.215407] env[62385]: DEBUG nova.virt.hardware [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 612.215566] env[62385]: DEBUG nova.virt.hardware [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 612.215752] env[62385]: DEBUG nova.virt.hardware [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 612.215883] env[62385]: DEBUG nova.virt.hardware [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 612.216068] env[62385]: DEBUG nova.virt.hardware [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 612.216951] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c97f75c9-cbcd-418b-84ae-7f5cb9e6fdba {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.225010] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77699217-6c85-42ea-b0b2-3f611ca7c090 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.239257] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Instance VIF info [] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 612.245236] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Creating folder: Project (c9ccdcaa831a4817be52284d9faeec76). Parent ref: group-v261107. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 612.245578] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1ba702c2-8903-4739-845b-34c2740fcc44 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.256708] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.087s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 612.257363] env[62385]: ERROR nova.compute.manager [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2e8c5043-214e-41e8-85e6-46fc8ecc3a4b, please check neutron logs for more information. [ 612.257363] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Traceback (most recent call last): [ 612.257363] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 612.257363] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] self.driver.spawn(context, instance, image_meta, [ 612.257363] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 612.257363] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] self._vmops.spawn(context, instance, image_meta, injected_files, [ 612.257363] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 612.257363] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] vm_ref = self.build_virtual_machine(instance, [ 612.257363] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 612.257363] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] vif_infos = vmwarevif.get_vif_info(self._session, [ 612.257363] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 612.257683] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] for vif in network_info: [ 612.257683] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 612.257683] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] return self._sync_wrapper(fn, *args, **kwargs) [ 612.257683] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 612.257683] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] self.wait() [ 612.257683] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 612.257683] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] self[:] = self._gt.wait() [ 612.257683] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 612.257683] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] return self._exit_event.wait() [ 612.257683] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 612.257683] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] current.throw(*self._exc) [ 612.257683] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 612.257683] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] result = function(*args, **kwargs) [ 612.258069] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 612.258069] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] return func(*args, **kwargs) [ 612.258069] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 612.258069] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] raise e [ 612.258069] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 612.258069] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] nwinfo = self.network_api.allocate_for_instance( [ 612.258069] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 612.258069] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] created_port_ids = self._update_ports_for_instance( [ 612.258069] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 612.258069] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] with excutils.save_and_reraise_exception(): [ 612.258069] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 612.258069] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] self.force_reraise() [ 612.258069] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 612.258524] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] raise self.value [ 612.258524] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 612.258524] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] updated_port = self._update_port( [ 612.258524] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 612.258524] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] _ensure_no_port_binding_failure(port) [ 612.258524] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 612.258524] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] raise exception.PortBindingFailed(port_id=port['id']) [ 612.258524] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] nova.exception.PortBindingFailed: Binding failed for port 2e8c5043-214e-41e8-85e6-46fc8ecc3a4b, please check neutron logs for more information. [ 612.258524] env[62385]: ERROR nova.compute.manager [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] [ 612.258524] env[62385]: DEBUG nova.compute.utils [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Binding failed for port 2e8c5043-214e-41e8-85e6-46fc8ecc3a4b, please check neutron logs for more information. {{(pid=62385) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 612.259363] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Created folder: Project (c9ccdcaa831a4817be52284d9faeec76) in parent group-v261107. [ 612.259539] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Creating folder: Instances. Parent ref: group-v261115. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 612.260061] env[62385]: DEBUG nova.compute.manager [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Build of instance 0a21307b-74e7-4c5d-95a3-0800915c51ff was re-scheduled: Binding failed for port 2e8c5043-214e-41e8-85e6-46fc8ecc3a4b, please check neutron logs for more information. {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 612.260813] env[62385]: DEBUG nova.compute.manager [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Unplugging VIFs for instance {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 612.260947] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] Acquiring lock "refresh_cache-0a21307b-74e7-4c5d-95a3-0800915c51ff" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.261107] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] Acquired lock "refresh_cache-0a21307b-74e7-4c5d-95a3-0800915c51ff" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.261269] env[62385]: DEBUG nova.network.neutron [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 612.262341] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.019s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.263715] env[62385]: INFO nova.compute.claims [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 612.266027] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6b2b0df6-0d10-4df0-b249-bcec846db999 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.275330] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Created folder: Instances in parent group-v261115. [ 612.275768] env[62385]: DEBUG oslo.service.loopingcall [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 612.275768] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 612.275862] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2e49babf-3390-41f2-9651-bea72d867ac2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.292113] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 612.292113] env[62385]: value = "task-1205450" [ 612.292113] env[62385]: _type = "Task" [ 612.292113] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.299922] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205450, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.465236] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] Releasing lock "refresh_cache-317e7e7c-7c6f-4855-a433-e55ea5571b92" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.466175] env[62385]: DEBUG nova.compute.manager [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 612.466529] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-330dbad4-16b1-4a5f-8511-81112ccbc057 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.475298] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffafe85b-92a1-46e4-9f9a-4b0e22965d1d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.509169] env[62385]: WARNING nova.virt.vmwareapi.driver [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 317e7e7c-7c6f-4855-a433-e55ea5571b92 could not be found. [ 612.509169] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 612.509169] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c76f1bc5-f934-4e69-a347-3f1fe35e2d28 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.512146] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df886879-a608-4d6b-a0f8-8d76af39390c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.536249] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 317e7e7c-7c6f-4855-a433-e55ea5571b92 could not be found. [ 612.536604] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 612.536714] env[62385]: INFO nova.compute.manager [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Took 0.07 seconds to destroy the instance on the hypervisor. [ 612.536935] env[62385]: DEBUG oslo.service.loopingcall [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 612.537184] env[62385]: DEBUG nova.compute.manager [-] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 612.537281] env[62385]: DEBUG nova.network.neutron [-] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 612.554333] env[62385]: DEBUG nova.network.neutron [-] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 612.793325] env[62385]: DEBUG nova.network.neutron [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 612.804393] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205450, 'name': CreateVM_Task, 'duration_secs': 0.251144} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.804439] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 612.805299] env[62385]: DEBUG oslo_vmware.service [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-522d0bb3-1268-483e-ae5a-13dc8762f3cb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.810827] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.810987] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.811432] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 612.812524] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d01d1287-6392-4a13-8f4f-b6790a2533c2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.818216] env[62385]: DEBUG oslo_vmware.api [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Waiting for the task: (returnval){ [ 612.818216] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e3d6ff-2e41-ac6b-3de4-cf4b09a27b65" [ 612.818216] env[62385]: _type = "Task" [ 612.818216] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.824888] env[62385]: DEBUG oslo_vmware.api [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e3d6ff-2e41-ac6b-3de4-cf4b09a27b65, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.889540] env[62385]: DEBUG nova.network.neutron [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.056938] env[62385]: DEBUG nova.network.neutron [-] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.330454] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.331284] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 613.331284] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 613.331284] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 613.331284] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 613.331471] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-baaafe87-1627-411b-8493-e1341a8a5473 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.352494] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 613.353076] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 613.353486] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8c39de6-62af-4717-9bf5-896561579619 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.362014] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dd244160-83a2-4b20-aa6f-f4f242c82227 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.366974] env[62385]: DEBUG oslo_vmware.api [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Waiting for the task: (returnval){ [ 613.366974] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ec04b3-57ce-c1f9-d2a8-6a9816fb1edf" [ 613.366974] env[62385]: _type = "Task" [ 613.366974] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.377613] env[62385]: DEBUG oslo_vmware.api [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ec04b3-57ce-c1f9-d2a8-6a9816fb1edf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.391720] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] Releasing lock "refresh_cache-0a21307b-74e7-4c5d-95a3-0800915c51ff" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.391943] env[62385]: DEBUG nova.compute.manager [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 613.392192] env[62385]: DEBUG nova.compute.manager [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 613.392317] env[62385]: DEBUG nova.network.neutron [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 613.406994] env[62385]: DEBUG nova.network.neutron [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 613.559031] env[62385]: INFO nova.compute.manager [-] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Took 1.02 seconds to deallocate network for instance. [ 613.667332] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb71bddb-9ec5-47ef-8668-0a8827227e77 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.674731] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc3f3eb7-c338-46b5-b9d0-b78262bf0aeb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.705072] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3750a60f-cb8b-451e-acf5-12ab7093235b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.712171] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f94d2eb-160e-41a8-a5ce-30d4964f5ea1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.728110] env[62385]: DEBUG nova.compute.provider_tree [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 613.880745] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Preparing fetch location {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 613.881038] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Creating directory with path [datastore2] vmware_temp/4a39aa97-49e9-4dc6-9846-40368890b5ad/c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 613.881292] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cc86d01b-1f28-4a8c-9b48-34c607fa14aa {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.904092] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Created directory with path [datastore2] vmware_temp/4a39aa97-49e9-4dc6-9846-40368890b5ad/c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 613.904324] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Fetch image to [datastore2] vmware_temp/4a39aa97-49e9-4dc6-9846-40368890b5ad/c0abbb8d-77e7-4f22-a256-0faf0a781109/tmp-sparse.vmdk {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 613.904488] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Downloading image file data c0abbb8d-77e7-4f22-a256-0faf0a781109 to [datastore2] vmware_temp/4a39aa97-49e9-4dc6-9846-40368890b5ad/c0abbb8d-77e7-4f22-a256-0faf0a781109/tmp-sparse.vmdk on the data store datastore2 {{(pid=62385) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 613.905321] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed95f9b9-df19-4ce2-b7b8-a6c00db7cf39 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.909744] env[62385]: DEBUG nova.network.neutron [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.914262] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2c188c6-9b3e-416a-b786-037b1a703404 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.924020] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fa326a5-415d-4244-9e9a-538c2b986911 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.955704] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0be526a-ff94-401e-bebf-698432b8f3ad {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.961752] env[62385]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-75c7a669-54ec-45d5-8e8b-55265efd08ec {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.050492] env[62385]: DEBUG nova.virt.vmwareapi.images [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Downloading image file data c0abbb8d-77e7-4f22-a256-0faf0a781109 to the data store datastore2 {{(pid=62385) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 614.113135] env[62385]: DEBUG oslo_vmware.rw_handles [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/4a39aa97-49e9-4dc6-9846-40368890b5ad/c0abbb8d-77e7-4f22-a256-0faf0a781109/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62385) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 614.173434] env[62385]: INFO nova.compute.manager [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Took 0.61 seconds to detach 1 volumes for instance. [ 614.175941] env[62385]: DEBUG nova.compute.claims [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Aborting claim: {{(pid=62385) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 614.176155] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.229610] env[62385]: DEBUG nova.scheduler.client.report [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 614.416126] env[62385]: INFO nova.compute.manager [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] [instance: 0a21307b-74e7-4c5d-95a3-0800915c51ff] Took 1.02 seconds to deallocate network for instance. [ 614.741946] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.479s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 614.742511] env[62385]: DEBUG nova.compute.manager [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 614.745529] env[62385]: DEBUG oslo_concurrency.lockutils [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.392s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.746936] env[62385]: INFO nova.compute.claims [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 614.845919] env[62385]: DEBUG oslo_vmware.rw_handles [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Completed reading data from the image iterator. {{(pid=62385) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 614.846162] env[62385]: DEBUG oslo_vmware.rw_handles [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Closing write handle for https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/4a39aa97-49e9-4dc6-9846-40368890b5ad/c0abbb8d-77e7-4f22-a256-0faf0a781109/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 614.896933] env[62385]: DEBUG nova.virt.vmwareapi.images [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Downloaded image file data c0abbb8d-77e7-4f22-a256-0faf0a781109 to vmware_temp/4a39aa97-49e9-4dc6-9846-40368890b5ad/c0abbb8d-77e7-4f22-a256-0faf0a781109/tmp-sparse.vmdk on the data store datastore2 {{(pid=62385) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 614.899061] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Caching image {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 614.899339] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Copying Virtual Disk [datastore2] vmware_temp/4a39aa97-49e9-4dc6-9846-40368890b5ad/c0abbb8d-77e7-4f22-a256-0faf0a781109/tmp-sparse.vmdk to [datastore2] vmware_temp/4a39aa97-49e9-4dc6-9846-40368890b5ad/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 614.900536] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2bb4c411-8fdc-4154-9050-38ec5188b9ab {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.908521] env[62385]: DEBUG oslo_vmware.api [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Waiting for the task: (returnval){ [ 614.908521] env[62385]: value = "task-1205451" [ 614.908521] env[62385]: _type = "Task" [ 614.908521] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.916687] env[62385]: DEBUG oslo_vmware.api [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205451, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.253490] env[62385]: DEBUG nova.compute.utils [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 615.257460] env[62385]: DEBUG nova.compute.manager [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Not allocating networking since 'none' was specified. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 615.418895] env[62385]: DEBUG oslo_vmware.api [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205451, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.452583] env[62385]: INFO nova.scheduler.client.report [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] Deleted allocations for instance 0a21307b-74e7-4c5d-95a3-0800915c51ff [ 615.758505] env[62385]: DEBUG nova.compute.manager [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 615.923212] env[62385]: DEBUG oslo_vmware.api [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205451, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.632545} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.928624] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Copied Virtual Disk [datastore2] vmware_temp/4a39aa97-49e9-4dc6-9846-40368890b5ad/c0abbb8d-77e7-4f22-a256-0faf0a781109/tmp-sparse.vmdk to [datastore2] vmware_temp/4a39aa97-49e9-4dc6-9846-40368890b5ad/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 615.928843] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Deleting the datastore file [datastore2] vmware_temp/4a39aa97-49e9-4dc6-9846-40368890b5ad/c0abbb8d-77e7-4f22-a256-0faf0a781109/tmp-sparse.vmdk {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 615.929358] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a219c73d-890b-418a-bca8-299f6720259e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.936697] env[62385]: DEBUG oslo_vmware.api [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Waiting for the task: (returnval){ [ 615.936697] env[62385]: value = "task-1205452" [ 615.936697] env[62385]: _type = "Task" [ 615.936697] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.946278] env[62385]: DEBUG oslo_vmware.api [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205452, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.961731] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8959f8e4-62df-4f98-b13a-46e19cd4c96f tempest-FloatingIPsAssociationTestJSON-1742054563 tempest-FloatingIPsAssociationTestJSON-1742054563-project-member] Lock "0a21307b-74e7-4c5d-95a3-0800915c51ff" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 79.228s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.221346] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0829b522-a140-4959-acb0-d8422a1b2720 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.228041] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7ccf20b-41d3-4275-a8e0-94b106fdc275 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.258295] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6a5eb18-025e-4eea-9ac4-8bdae0a124dc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.266019] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70874db1-b3a3-4eb6-a866-62f97c111c0d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.283526] env[62385]: DEBUG nova.compute.provider_tree [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 616.446891] env[62385]: DEBUG oslo_vmware.api [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205452, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.022674} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.447166] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 616.447399] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Moving file from [datastore2] vmware_temp/4a39aa97-49e9-4dc6-9846-40368890b5ad/c0abbb8d-77e7-4f22-a256-0faf0a781109 to [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109. {{(pid=62385) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 616.447691] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-b36b854a-7b85-4fd3-861b-3ff6fb67b0d9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.454039] env[62385]: DEBUG oslo_vmware.api [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Waiting for the task: (returnval){ [ 616.454039] env[62385]: value = "task-1205453" [ 616.454039] env[62385]: _type = "Task" [ 616.454039] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.461458] env[62385]: DEBUG oslo_vmware.api [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205453, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.464217] env[62385]: DEBUG nova.compute.manager [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] [instance: ad12e5b0-6854-4281-9155-e1660393829c] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 616.777324] env[62385]: DEBUG nova.compute.manager [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 616.786481] env[62385]: DEBUG nova.scheduler.client.report [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 616.802512] env[62385]: DEBUG nova.virt.hardware [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 616.802774] env[62385]: DEBUG nova.virt.hardware [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 616.802942] env[62385]: DEBUG nova.virt.hardware [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 616.803138] env[62385]: DEBUG nova.virt.hardware [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 616.803285] env[62385]: DEBUG nova.virt.hardware [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 616.803429] env[62385]: DEBUG nova.virt.hardware [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 616.803635] env[62385]: DEBUG nova.virt.hardware [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 616.803793] env[62385]: DEBUG nova.virt.hardware [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 616.803957] env[62385]: DEBUG nova.virt.hardware [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 616.804209] env[62385]: DEBUG nova.virt.hardware [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 616.804398] env[62385]: DEBUG nova.virt.hardware [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 616.805287] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed595d4a-1eb7-4fef-b4e2-5494f28d13ed {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.816019] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-820e03e2-cc3f-4c4e-8b67-c11164742b00 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.829340] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Instance VIF info [] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 616.836741] env[62385]: DEBUG oslo.service.loopingcall [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 616.837231] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 616.837456] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5914f206-6fed-4c62-ac82-cc7ce4d6c63b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.854723] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 616.854723] env[62385]: value = "task-1205454" [ 616.854723] env[62385]: _type = "Task" [ 616.854723] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.861948] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205454, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.967764] env[62385]: DEBUG oslo_vmware.api [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205453, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.026061} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.968156] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] File moved {{(pid=62385) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 616.968779] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Cleaning up location [datastore2] vmware_temp/4a39aa97-49e9-4dc6-9846-40368890b5ad {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 616.969322] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Deleting the datastore file [datastore2] vmware_temp/4a39aa97-49e9-4dc6-9846-40368890b5ad {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 616.972774] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-32ba2229-14d1-4bc8-a9e9-be6843c57bf5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.982844] env[62385]: DEBUG oslo_vmware.api [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Waiting for the task: (returnval){ [ 616.982844] env[62385]: value = "task-1205455" [ 616.982844] env[62385]: _type = "Task" [ 616.982844] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.991981] env[62385]: DEBUG oslo_vmware.api [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205455, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.993980] env[62385]: DEBUG oslo_concurrency.lockutils [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.291394] env[62385]: DEBUG oslo_concurrency.lockutils [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.546s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 617.291928] env[62385]: DEBUG nova.compute.manager [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 617.294569] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.088s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.365709] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205454, 'name': CreateVM_Task, 'duration_secs': 0.307261} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.365875] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 617.366836] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 617.367028] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.367322] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 617.368072] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9255a417-f668-4465-8236-b906156237bc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.373058] env[62385]: DEBUG oslo_vmware.api [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Waiting for the task: (returnval){ [ 617.373058] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52b05b7e-92c2-f387-18c1-38f1bc2b09c5" [ 617.373058] env[62385]: _type = "Task" [ 617.373058] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.382810] env[62385]: DEBUG oslo_vmware.api [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52b05b7e-92c2-f387-18c1-38f1bc2b09c5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.493044] env[62385]: DEBUG oslo_vmware.api [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205455, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.024796} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.493120] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 617.493840] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bed8cfac-1afc-4ed0-a759-f0376773453f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.498852] env[62385]: DEBUG oslo_vmware.api [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Waiting for the task: (returnval){ [ 617.498852] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521e470f-32a7-03c9-caaf-309a55f7fe67" [ 617.498852] env[62385]: _type = "Task" [ 617.498852] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.506818] env[62385]: DEBUG oslo_vmware.api [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521e470f-32a7-03c9-caaf-309a55f7fe67, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.800164] env[62385]: DEBUG nova.compute.utils [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 617.808191] env[62385]: DEBUG nova.compute.manager [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 617.811177] env[62385]: DEBUG nova.network.neutron [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 617.853785] env[62385]: DEBUG nova.policy [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e1200cdf75b04233b4c3f4626d4e27cf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aea48968d4464ac0ab01c998c059107d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 617.886338] env[62385]: DEBUG oslo_vmware.api [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52b05b7e-92c2-f387-18c1-38f1bc2b09c5, 'name': SearchDatastore_Task, 'duration_secs': 0.009213} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.886684] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 617.886969] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 617.887530] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 617.887530] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.887710] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 617.888460] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e414fe86-6484-47de-a30f-fa5c1aa7743f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.896209] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 617.897148] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 617.899610] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2468f4d3-5127-4f16-ad80-cfe86cda272c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.905471] env[62385]: DEBUG oslo_vmware.api [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Waiting for the task: (returnval){ [ 617.905471] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a347be-dc72-7efc-51f0-2a2505215e36" [ 617.905471] env[62385]: _type = "Task" [ 617.905471] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.917095] env[62385]: DEBUG oslo_vmware.api [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a347be-dc72-7efc-51f0-2a2505215e36, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.012910] env[62385]: DEBUG oslo_vmware.api [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521e470f-32a7-03c9-caaf-309a55f7fe67, 'name': SearchDatastore_Task, 'duration_secs': 0.008351} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.013195] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 618.013465] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 0ba32627-c646-477d-9062-b84624d01e48/0ba32627-c646-477d-9062-b84624d01e48.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 618.014027] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0ddc693a-bcaa-4b91-9efe-14de9854a51c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.024019] env[62385]: DEBUG oslo_vmware.api [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Waiting for the task: (returnval){ [ 618.024019] env[62385]: value = "task-1205456" [ 618.024019] env[62385]: _type = "Task" [ 618.024019] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.031201] env[62385]: DEBUG oslo_vmware.api [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205456, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.178470] env[62385]: DEBUG nova.network.neutron [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Successfully created port: 6eba3a70-eb34-4ef7-8145-12a451cc108b {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 618.309434] env[62385]: DEBUG nova.compute.manager [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 618.379603] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-733d19bf-de5e-40bb-afee-c960d3539ce6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.390539] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93467f53-b609-4c15-8602-cfcf14984a59 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.426178] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fed9c11-c7cb-4b29-91cc-1be5f60567a3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.439406] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff737184-dfd9-4164-bb67-ffae704fdde1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.443538] env[62385]: DEBUG oslo_vmware.api [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a347be-dc72-7efc-51f0-2a2505215e36, 'name': SearchDatastore_Task, 'duration_secs': 0.008806} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.444947] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66a782e1-317e-46e4-8f6e-909b5405977d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.455555] env[62385]: DEBUG nova.compute.provider_tree [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 618.460065] env[62385]: DEBUG oslo_vmware.api [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Waiting for the task: (returnval){ [ 618.460065] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52356326-0ed2-3c34-486c-e272a3b9895d" [ 618.460065] env[62385]: _type = "Task" [ 618.460065] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.467758] env[62385]: DEBUG oslo_vmware.api [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52356326-0ed2-3c34-486c-e272a3b9895d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.533179] env[62385]: DEBUG oslo_vmware.api [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205456, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.961359] env[62385]: DEBUG nova.scheduler.client.report [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 618.972128] env[62385]: DEBUG oslo_vmware.api [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52356326-0ed2-3c34-486c-e272a3b9895d, 'name': SearchDatastore_Task, 'duration_secs': 0.051577} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.972640] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 618.973014] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] f0f0eed0-a76a-42bc-9cca-91b93d7ad64c/f0f0eed0-a76a-42bc-9cca-91b93d7ad64c.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 618.973403] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8e37f0ef-a183-4273-8754-f9657f8e4bea {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.982126] env[62385]: DEBUG oslo_vmware.api [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Waiting for the task: (returnval){ [ 618.982126] env[62385]: value = "task-1205457" [ 618.982126] env[62385]: _type = "Task" [ 618.982126] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.989652] env[62385]: DEBUG oslo_vmware.api [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205457, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.034203] env[62385]: DEBUG oslo_vmware.api [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205456, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.520797} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.034203] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 0ba32627-c646-477d-9062-b84624d01e48/0ba32627-c646-477d-9062-b84624d01e48.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 619.034608] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 619.034967] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cf696fae-a0f2-4f23-b62d-4692a7ff7fba {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.042393] env[62385]: DEBUG oslo_vmware.api [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Waiting for the task: (returnval){ [ 619.042393] env[62385]: value = "task-1205458" [ 619.042393] env[62385]: _type = "Task" [ 619.042393] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.052936] env[62385]: DEBUG oslo_vmware.api [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205458, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.319370] env[62385]: DEBUG nova.compute.manager [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 619.354889] env[62385]: DEBUG nova.virt.hardware [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 619.355048] env[62385]: DEBUG nova.virt.hardware [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 619.355142] env[62385]: DEBUG nova.virt.hardware [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 619.356053] env[62385]: DEBUG nova.virt.hardware [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 619.356053] env[62385]: DEBUG nova.virt.hardware [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 619.356053] env[62385]: DEBUG nova.virt.hardware [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 619.356053] env[62385]: DEBUG nova.virt.hardware [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 619.356053] env[62385]: DEBUG nova.virt.hardware [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 619.356274] env[62385]: DEBUG nova.virt.hardware [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 619.356364] env[62385]: DEBUG nova.virt.hardware [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 619.356587] env[62385]: DEBUG nova.virt.hardware [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 619.357510] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cad0f81-17b2-44bc-a63e-888b38616198 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.363284] env[62385]: DEBUG nova.compute.manager [req-b6296716-177b-4c34-88ff-d41130a6fa27 req-a4c5c2e9-2387-40f2-b2c9-4b691265515d service nova] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Received event network-changed-6eba3a70-eb34-4ef7-8145-12a451cc108b {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 619.363478] env[62385]: DEBUG nova.compute.manager [req-b6296716-177b-4c34-88ff-d41130a6fa27 req-a4c5c2e9-2387-40f2-b2c9-4b691265515d service nova] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Refreshing instance network info cache due to event network-changed-6eba3a70-eb34-4ef7-8145-12a451cc108b. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 619.364171] env[62385]: DEBUG oslo_concurrency.lockutils [req-b6296716-177b-4c34-88ff-d41130a6fa27 req-a4c5c2e9-2387-40f2-b2c9-4b691265515d service nova] Acquiring lock "refresh_cache-01d010dc-6b5b-470b-a387-f95a17600e19" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 619.364338] env[62385]: DEBUG oslo_concurrency.lockutils [req-b6296716-177b-4c34-88ff-d41130a6fa27 req-a4c5c2e9-2387-40f2-b2c9-4b691265515d service nova] Acquired lock "refresh_cache-01d010dc-6b5b-470b-a387-f95a17600e19" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.364507] env[62385]: DEBUG nova.network.neutron [req-b6296716-177b-4c34-88ff-d41130a6fa27 req-a4c5c2e9-2387-40f2-b2c9-4b691265515d service nova] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Refreshing network info cache for port 6eba3a70-eb34-4ef7-8145-12a451cc108b {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 619.369590] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29c54bdd-76fe-47d8-88ea-8051cac6e9b3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.466927] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.172s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.467629] env[62385]: ERROR nova.compute.manager [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e057457e-a781-4fd0-b8da-a1ea032ef872, please check neutron logs for more information. [ 619.467629] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Traceback (most recent call last): [ 619.467629] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 619.467629] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] self.driver.spawn(context, instance, image_meta, [ 619.467629] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 619.467629] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 619.467629] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 619.467629] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] vm_ref = self.build_virtual_machine(instance, [ 619.467629] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 619.467629] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] vif_infos = vmwarevif.get_vif_info(self._session, [ 619.467629] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 619.468711] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] for vif in network_info: [ 619.468711] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 619.468711] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] return self._sync_wrapper(fn, *args, **kwargs) [ 619.468711] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 619.468711] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] self.wait() [ 619.468711] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 619.468711] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] self[:] = self._gt.wait() [ 619.468711] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 619.468711] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] return self._exit_event.wait() [ 619.468711] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 619.468711] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] current.throw(*self._exc) [ 619.468711] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 619.468711] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] result = function(*args, **kwargs) [ 619.469090] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 619.469090] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] return func(*args, **kwargs) [ 619.469090] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 619.469090] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] raise e [ 619.469090] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 619.469090] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] nwinfo = self.network_api.allocate_for_instance( [ 619.469090] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 619.469090] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] created_port_ids = self._update_ports_for_instance( [ 619.469090] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 619.469090] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] with excutils.save_and_reraise_exception(): [ 619.469090] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 619.469090] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] self.force_reraise() [ 619.469090] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 619.469441] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] raise self.value [ 619.469441] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 619.469441] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] updated_port = self._update_port( [ 619.469441] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 619.469441] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] _ensure_no_port_binding_failure(port) [ 619.469441] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 619.469441] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] raise exception.PortBindingFailed(port_id=port['id']) [ 619.469441] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] nova.exception.PortBindingFailed: Binding failed for port e057457e-a781-4fd0-b8da-a1ea032ef872, please check neutron logs for more information. [ 619.469441] env[62385]: ERROR nova.compute.manager [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] [ 619.469441] env[62385]: DEBUG nova.compute.utils [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Binding failed for port e057457e-a781-4fd0-b8da-a1ea032ef872, please check neutron logs for more information. {{(pid=62385) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 619.469783] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 18.282s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.470327] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.470327] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62385) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 619.470327] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.319s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.472248] env[62385]: INFO nova.compute.claims [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 619.475087] env[62385]: DEBUG nova.compute.manager [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Build of instance ce6b8f31-afee-4062-b860-d054c4d37be0 was re-scheduled: Binding failed for port e057457e-a781-4fd0-b8da-a1ea032ef872, please check neutron logs for more information. {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 619.475536] env[62385]: DEBUG nova.compute.manager [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Unplugging VIFs for instance {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 619.475924] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "refresh_cache-ce6b8f31-afee-4062-b860-d054c4d37be0" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 619.475924] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquired lock "refresh_cache-ce6b8f31-afee-4062-b860-d054c4d37be0" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.476067] env[62385]: DEBUG nova.network.neutron [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 619.477483] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a36b0d40-3f91-46e0-bc10-a23f3a202e8f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.486954] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff6fd653-9057-41b2-bcaf-d01998d14dbe {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.494045] env[62385]: DEBUG oslo_vmware.api [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205457, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.462949} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.494628] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] f0f0eed0-a76a-42bc-9cca-91b93d7ad64c/f0f0eed0-a76a-42bc-9cca-91b93d7ad64c.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 619.494823] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 619.495319] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d0c8b821-ecff-4222-97e6-29ca697fe14c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.508602] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e650dc67-268c-492a-97c3-878eaff7bc67 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.512647] env[62385]: DEBUG oslo_vmware.api [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Waiting for the task: (returnval){ [ 619.512647] env[62385]: value = "task-1205459" [ 619.512647] env[62385]: _type = "Task" [ 619.512647] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.520821] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91eea124-4c58-4f30-a7bf-56ea75dd70b7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.526154] env[62385]: DEBUG oslo_vmware.api [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205459, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.526999] env[62385]: ERROR nova.compute.manager [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6eba3a70-eb34-4ef7-8145-12a451cc108b, please check neutron logs for more information. [ 619.526999] env[62385]: ERROR nova.compute.manager Traceback (most recent call last): [ 619.526999] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 619.526999] env[62385]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 619.526999] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 619.526999] env[62385]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 619.526999] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 619.526999] env[62385]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 619.526999] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 619.526999] env[62385]: ERROR nova.compute.manager self.force_reraise() [ 619.526999] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 619.526999] env[62385]: ERROR nova.compute.manager raise self.value [ 619.526999] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 619.526999] env[62385]: ERROR nova.compute.manager updated_port = self._update_port( [ 619.526999] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 619.526999] env[62385]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 619.527436] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 619.527436] env[62385]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 619.527436] env[62385]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6eba3a70-eb34-4ef7-8145-12a451cc108b, please check neutron logs for more information. [ 619.527436] env[62385]: ERROR nova.compute.manager [ 619.527436] env[62385]: Traceback (most recent call last): [ 619.527436] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 619.527436] env[62385]: listener.cb(fileno) [ 619.527436] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 619.527436] env[62385]: result = function(*args, **kwargs) [ 619.527436] env[62385]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 619.527436] env[62385]: return func(*args, **kwargs) [ 619.527436] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 619.527436] env[62385]: raise e [ 619.527436] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 619.527436] env[62385]: nwinfo = self.network_api.allocate_for_instance( [ 619.527436] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 619.527436] env[62385]: created_port_ids = self._update_ports_for_instance( [ 619.527436] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 619.527436] env[62385]: with excutils.save_and_reraise_exception(): [ 619.527436] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 619.527436] env[62385]: self.force_reraise() [ 619.527436] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 619.527436] env[62385]: raise self.value [ 619.527436] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 619.527436] env[62385]: updated_port = self._update_port( [ 619.527436] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 619.527436] env[62385]: _ensure_no_port_binding_failure(port) [ 619.527436] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 619.527436] env[62385]: raise exception.PortBindingFailed(port_id=port['id']) [ 619.528124] env[62385]: nova.exception.PortBindingFailed: Binding failed for port 6eba3a70-eb34-4ef7-8145-12a451cc108b, please check neutron logs for more information. [ 619.528124] env[62385]: Removing descriptor: 17 [ 619.528333] env[62385]: ERROR nova.compute.manager [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6eba3a70-eb34-4ef7-8145-12a451cc108b, please check neutron logs for more information. [ 619.528333] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Traceback (most recent call last): [ 619.528333] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 619.528333] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] yield resources [ 619.528333] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 619.528333] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] self.driver.spawn(context, instance, image_meta, [ 619.528333] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 619.528333] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] self._vmops.spawn(context, instance, image_meta, injected_files, [ 619.528333] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 619.528333] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] vm_ref = self.build_virtual_machine(instance, [ 619.528333] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 619.528935] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] vif_infos = vmwarevif.get_vif_info(self._session, [ 619.528935] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 619.528935] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] for vif in network_info: [ 619.528935] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 619.528935] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] return self._sync_wrapper(fn, *args, **kwargs) [ 619.528935] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 619.528935] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] self.wait() [ 619.528935] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 619.528935] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] self[:] = self._gt.wait() [ 619.528935] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 619.528935] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] return self._exit_event.wait() [ 619.528935] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 619.528935] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] result = hub.switch() [ 619.529424] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 619.529424] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] return self.greenlet.switch() [ 619.529424] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 619.529424] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] result = function(*args, **kwargs) [ 619.529424] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 619.529424] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] return func(*args, **kwargs) [ 619.529424] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 619.529424] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] raise e [ 619.529424] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 619.529424] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] nwinfo = self.network_api.allocate_for_instance( [ 619.529424] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 619.529424] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] created_port_ids = self._update_ports_for_instance( [ 619.529424] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 619.530112] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] with excutils.save_and_reraise_exception(): [ 619.530112] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 619.530112] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] self.force_reraise() [ 619.530112] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 619.530112] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] raise self.value [ 619.530112] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 619.530112] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] updated_port = self._update_port( [ 619.530112] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 619.530112] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] _ensure_no_port_binding_failure(port) [ 619.530112] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 619.530112] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] raise exception.PortBindingFailed(port_id=port['id']) [ 619.530112] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] nova.exception.PortBindingFailed: Binding failed for port 6eba3a70-eb34-4ef7-8145-12a451cc108b, please check neutron logs for more information. [ 619.530112] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] [ 619.531141] env[62385]: INFO nova.compute.manager [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Terminating instance [ 619.555075] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181387MB free_disk=155GB free_vcpus=48 pci_devices=None {{(pid=62385) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 619.555187] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.555750] env[62385]: DEBUG oslo_concurrency.lockutils [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "refresh_cache-01d010dc-6b5b-470b-a387-f95a17600e19" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 619.563621] env[62385]: DEBUG oslo_vmware.api [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205458, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065585} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.564031] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 619.564647] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-314e5b83-bf72-40ce-b3a5-58e82ed49e65 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.584340] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Reconfiguring VM instance instance-00000018 to attach disk [datastore2] 0ba32627-c646-477d-9062-b84624d01e48/0ba32627-c646-477d-9062-b84624d01e48.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 619.584935] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9ff9d20a-2a0f-48ee-bbb7-ec65751fd1ea {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.607154] env[62385]: DEBUG oslo_vmware.api [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Waiting for the task: (returnval){ [ 619.607154] env[62385]: value = "task-1205460" [ 619.607154] env[62385]: _type = "Task" [ 619.607154] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.612945] env[62385]: DEBUG oslo_vmware.api [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205460, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.890810] env[62385]: DEBUG nova.network.neutron [req-b6296716-177b-4c34-88ff-d41130a6fa27 req-a4c5c2e9-2387-40f2-b2c9-4b691265515d service nova] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 620.009208] env[62385]: DEBUG nova.network.neutron [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 620.023164] env[62385]: DEBUG oslo_vmware.api [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205459, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.027729] env[62385]: DEBUG nova.network.neutron [req-b6296716-177b-4c34-88ff-d41130a6fa27 req-a4c5c2e9-2387-40f2-b2c9-4b691265515d service nova] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.117174] env[62385]: DEBUG oslo_vmware.api [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205460, 'name': ReconfigVM_Task, 'duration_secs': 0.277653} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.117174] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Reconfigured VM instance instance-00000018 to attach disk [datastore2] 0ba32627-c646-477d-9062-b84624d01e48/0ba32627-c646-477d-9062-b84624d01e48.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 620.117174] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5f03d078-9bda-45c2-9b27-5be36a9359b7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.123373] env[62385]: DEBUG oslo_vmware.api [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Waiting for the task: (returnval){ [ 620.123373] env[62385]: value = "task-1205461" [ 620.123373] env[62385]: _type = "Task" [ 620.123373] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.131254] env[62385]: DEBUG oslo_vmware.api [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205461, 'name': Rename_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.151476] env[62385]: DEBUG nova.network.neutron [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.523091] env[62385]: DEBUG oslo_vmware.api [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205459, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.989101} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.527917] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 620.528377] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37d009f9-4afa-4ae7-a753-9905197cbcfc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.531213] env[62385]: DEBUG oslo_concurrency.lockutils [req-b6296716-177b-4c34-88ff-d41130a6fa27 req-a4c5c2e9-2387-40f2-b2c9-4b691265515d service nova] Releasing lock "refresh_cache-01d010dc-6b5b-470b-a387-f95a17600e19" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 620.531691] env[62385]: DEBUG oslo_concurrency.lockutils [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquired lock "refresh_cache-01d010dc-6b5b-470b-a387-f95a17600e19" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 620.533431] env[62385]: DEBUG nova.network.neutron [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 620.557209] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Reconfiguring VM instance instance-00000019 to attach disk [datastore1] f0f0eed0-a76a-42bc-9cca-91b93d7ad64c/f0f0eed0-a76a-42bc-9cca-91b93d7ad64c.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 620.560709] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e1de1020-783d-4f03-bf97-49531185e810 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.583058] env[62385]: DEBUG oslo_vmware.api [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Waiting for the task: (returnval){ [ 620.583058] env[62385]: value = "task-1205462" [ 620.583058] env[62385]: _type = "Task" [ 620.583058] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.596179] env[62385]: DEBUG oslo_vmware.api [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205462, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.634927] env[62385]: DEBUG oslo_vmware.api [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205461, 'name': Rename_Task, 'duration_secs': 0.128965} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.635232] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 620.635477] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e68ea14e-c1d2-465c-ba99-3666664aac4c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.643940] env[62385]: DEBUG oslo_vmware.api [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Waiting for the task: (returnval){ [ 620.643940] env[62385]: value = "task-1205463" [ 620.643940] env[62385]: _type = "Task" [ 620.643940] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.652653] env[62385]: DEBUG oslo_vmware.api [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205463, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.657112] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Releasing lock "refresh_cache-ce6b8f31-afee-4062-b860-d054c4d37be0" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 620.657344] env[62385]: DEBUG nova.compute.manager [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 620.657555] env[62385]: DEBUG nova.compute.manager [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 620.657738] env[62385]: DEBUG nova.network.neutron [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 620.687524] env[62385]: DEBUG nova.network.neutron [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 621.015941] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e85f20a6-b08c-4448-8d8f-7ad669ba344c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.023505] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1319c22e-921b-4730-98e1-0ccb2bdeccfa {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.057960] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d69dd426-8729-401f-939e-80d2a8c293d6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.065953] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0521dc9c-2814-4047-bed8-2fad536e2807 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.080520] env[62385]: DEBUG nova.compute.provider_tree [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 621.089274] env[62385]: DEBUG nova.network.neutron [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 621.094282] env[62385]: DEBUG oslo_vmware.api [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205462, 'name': ReconfigVM_Task, 'duration_secs': 0.368321} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.095141] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Reconfigured VM instance instance-00000019 to attach disk [datastore1] f0f0eed0-a76a-42bc-9cca-91b93d7ad64c/f0f0eed0-a76a-42bc-9cca-91b93d7ad64c.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 621.095745] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bc6db5a9-d4f7-4e01-8ebe-a3f9462a87e1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.101647] env[62385]: DEBUG oslo_vmware.api [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Waiting for the task: (returnval){ [ 621.101647] env[62385]: value = "task-1205464" [ 621.101647] env[62385]: _type = "Task" [ 621.101647] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.110370] env[62385]: DEBUG oslo_vmware.api [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205464, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.153007] env[62385]: DEBUG oslo_vmware.api [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205463, 'name': PowerOnVM_Task, 'duration_secs': 0.44226} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.157504] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 621.157717] env[62385]: INFO nova.compute.manager [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Took 8.97 seconds to spawn the instance on the hypervisor. [ 621.157916] env[62385]: DEBUG nova.compute.manager [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 621.158765] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19b11c8e-c089-4ef4-aea8-3e5e6ce5d14b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.192561] env[62385]: DEBUG nova.network.neutron [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.231512] env[62385]: DEBUG nova.network.neutron [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.392881] env[62385]: DEBUG nova.compute.manager [req-0c923329-8482-44b8-936c-d6a033d2b5cd req-d12c3af7-7c0f-473d-899f-fc613470d378 service nova] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Received event network-vif-deleted-6eba3a70-eb34-4ef7-8145-12a451cc108b {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 621.585908] env[62385]: DEBUG nova.scheduler.client.report [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 621.613736] env[62385]: DEBUG oslo_vmware.api [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205464, 'name': Rename_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.677350] env[62385]: INFO nova.compute.manager [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Took 30.36 seconds to build instance. [ 621.696127] env[62385]: INFO nova.compute.manager [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: ce6b8f31-afee-4062-b860-d054c4d37be0] Took 1.04 seconds to deallocate network for instance. [ 621.736084] env[62385]: DEBUG oslo_concurrency.lockutils [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Releasing lock "refresh_cache-01d010dc-6b5b-470b-a387-f95a17600e19" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 621.736500] env[62385]: DEBUG nova.compute.manager [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 621.736706] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 621.737242] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-adcf7e2f-d788-46d7-9086-daa890f463dc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.747666] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fce7f94f-e080-4839-b6d1-32c797c328ff {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.771491] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 01d010dc-6b5b-470b-a387-f95a17600e19 could not be found. [ 621.772262] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 621.772262] env[62385]: INFO nova.compute.manager [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Took 0.04 seconds to destroy the instance on the hypervisor. [ 621.772378] env[62385]: DEBUG oslo.service.loopingcall [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 621.772531] env[62385]: DEBUG nova.compute.manager [-] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 621.772671] env[62385]: DEBUG nova.network.neutron [-] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 621.787462] env[62385]: DEBUG nova.network.neutron [-] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 622.091389] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.621s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.091857] env[62385]: DEBUG nova.compute.manager [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 622.094582] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.692s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.096042] env[62385]: INFO nova.compute.claims [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 622.111896] env[62385]: DEBUG oslo_vmware.api [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205464, 'name': Rename_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.179498] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fbc1f57a-6198-4256-b51d-90d27e1f1f70 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Lock "0ba32627-c646-477d-9062-b84624d01e48" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.486s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.289893] env[62385]: DEBUG nova.network.neutron [-] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.600529] env[62385]: DEBUG nova.compute.utils [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 622.603635] env[62385]: DEBUG nova.compute.manager [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 622.603804] env[62385]: DEBUG nova.network.neutron [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 622.615461] env[62385]: DEBUG oslo_vmware.api [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205464, 'name': Rename_Task, 'duration_secs': 1.348595} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.615704] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 622.615931] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9e68e941-2f96-44ab-8c9b-2fe61fbaca34 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.622859] env[62385]: DEBUG oslo_vmware.api [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Waiting for the task: (returnval){ [ 622.622859] env[62385]: value = "task-1205465" [ 622.622859] env[62385]: _type = "Task" [ 622.622859] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.631795] env[62385]: DEBUG oslo_vmware.api [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205465, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.683656] env[62385]: DEBUG nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: da94a435-b36c-42b2-8c84-564265872048] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 622.688214] env[62385]: DEBUG nova.policy [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e85ad9e3f4d14f80a246bdb1069b652b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '079d069fcb7446bfafc1fdcf283b27ab', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 622.726916] env[62385]: INFO nova.scheduler.client.report [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Deleted allocations for instance ce6b8f31-afee-4062-b860-d054c4d37be0 [ 622.792623] env[62385]: INFO nova.compute.manager [-] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Took 1.02 seconds to deallocate network for instance. [ 622.795535] env[62385]: DEBUG nova.compute.claims [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Aborting claim: {{(pid=62385) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 622.795850] env[62385]: DEBUG oslo_concurrency.lockutils [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.109755] env[62385]: DEBUG nova.compute.manager [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 623.115882] env[62385]: DEBUG nova.network.neutron [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Successfully created port: baf9e6b1-3e80-4e7f-9ed6-e6cdabd7320a {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 623.132107] env[62385]: DEBUG oslo_vmware.api [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205465, 'name': PowerOnVM_Task, 'duration_secs': 0.449998} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.135056] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 623.135283] env[62385]: INFO nova.compute.manager [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Took 6.36 seconds to spawn the instance on the hypervisor. [ 623.135463] env[62385]: DEBUG nova.compute.manager [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 623.136486] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fcdb95c-cf14-4c97-a3eb-637070e6106d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.208379] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.235296] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c7cb493-b0c0-4ff9-b447-813e4f4f193e tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "ce6b8f31-afee-4062-b860-d054c4d37be0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 85.667s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.631385] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-455d40ec-928a-443e-a052-6b3f636dfccc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.642626] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f06a5844-2006-4d43-9164-f214e8fc478c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.678189] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1954ab0-9b27-44a7-ab9e-bdb1fe96f5ac {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.687292] env[62385]: INFO nova.compute.manager [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Took 28.46 seconds to build instance. [ 623.693775] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43ee2eba-09a3-4932-882f-1a490a6bdac4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.709213] env[62385]: DEBUG nova.compute.provider_tree [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 623.737517] env[62385]: DEBUG nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 624.122224] env[62385]: DEBUG nova.compute.manager [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 624.126982] env[62385]: INFO nova.compute.manager [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Rebuilding instance [ 624.153949] env[62385]: DEBUG nova.virt.hardware [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 624.154225] env[62385]: DEBUG nova.virt.hardware [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 624.154388] env[62385]: DEBUG nova.virt.hardware [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 624.154568] env[62385]: DEBUG nova.virt.hardware [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 624.154711] env[62385]: DEBUG nova.virt.hardware [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 624.154854] env[62385]: DEBUG nova.virt.hardware [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 624.155288] env[62385]: DEBUG nova.virt.hardware [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 624.155563] env[62385]: DEBUG nova.virt.hardware [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 624.155808] env[62385]: DEBUG nova.virt.hardware [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 624.156045] env[62385]: DEBUG nova.virt.hardware [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 624.156285] env[62385]: DEBUG nova.virt.hardware [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 624.158382] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd485756-1c33-4d66-8232-cb143d3bb528 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.171575] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f903f55-d236-4577-8b19-75921938a087 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.193187] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8e56efe8-8044-4943-a780-7f4054fcd227 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Lock "f0f0eed0-a76a-42bc-9cca-91b93d7ad64c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.734s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 624.193677] env[62385]: DEBUG nova.compute.manager [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 624.195272] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8b929c7-9c42-43f8-a19f-47d9f3b19d85 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.212474] env[62385]: DEBUG nova.scheduler.client.report [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 624.257309] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.467398] env[62385]: ERROR nova.compute.manager [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port baf9e6b1-3e80-4e7f-9ed6-e6cdabd7320a, please check neutron logs for more information. [ 624.467398] env[62385]: ERROR nova.compute.manager Traceback (most recent call last): [ 624.467398] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.467398] env[62385]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 624.467398] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 624.467398] env[62385]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 624.467398] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 624.467398] env[62385]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 624.467398] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.467398] env[62385]: ERROR nova.compute.manager self.force_reraise() [ 624.467398] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.467398] env[62385]: ERROR nova.compute.manager raise self.value [ 624.467398] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 624.467398] env[62385]: ERROR nova.compute.manager updated_port = self._update_port( [ 624.467398] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.467398] env[62385]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 624.467842] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.467842] env[62385]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 624.467842] env[62385]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port baf9e6b1-3e80-4e7f-9ed6-e6cdabd7320a, please check neutron logs for more information. [ 624.467842] env[62385]: ERROR nova.compute.manager [ 624.467842] env[62385]: Traceback (most recent call last): [ 624.467842] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 624.467842] env[62385]: listener.cb(fileno) [ 624.467842] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 624.467842] env[62385]: result = function(*args, **kwargs) [ 624.467842] env[62385]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 624.467842] env[62385]: return func(*args, **kwargs) [ 624.467842] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 624.467842] env[62385]: raise e [ 624.467842] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.467842] env[62385]: nwinfo = self.network_api.allocate_for_instance( [ 624.467842] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 624.467842] env[62385]: created_port_ids = self._update_ports_for_instance( [ 624.467842] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 624.467842] env[62385]: with excutils.save_and_reraise_exception(): [ 624.467842] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.467842] env[62385]: self.force_reraise() [ 624.467842] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.467842] env[62385]: raise self.value [ 624.467842] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 624.467842] env[62385]: updated_port = self._update_port( [ 624.467842] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.467842] env[62385]: _ensure_no_port_binding_failure(port) [ 624.467842] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.467842] env[62385]: raise exception.PortBindingFailed(port_id=port['id']) [ 624.468838] env[62385]: nova.exception.PortBindingFailed: Binding failed for port baf9e6b1-3e80-4e7f-9ed6-e6cdabd7320a, please check neutron logs for more information. [ 624.468838] env[62385]: Removing descriptor: 19 [ 624.468838] env[62385]: ERROR nova.compute.manager [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port baf9e6b1-3e80-4e7f-9ed6-e6cdabd7320a, please check neutron logs for more information. [ 624.468838] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Traceback (most recent call last): [ 624.468838] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 624.468838] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] yield resources [ 624.468838] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 624.468838] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] self.driver.spawn(context, instance, image_meta, [ 624.468838] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 624.468838] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 624.468838] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 624.468838] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] vm_ref = self.build_virtual_machine(instance, [ 624.469348] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 624.469348] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] vif_infos = vmwarevif.get_vif_info(self._session, [ 624.469348] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 624.469348] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] for vif in network_info: [ 624.469348] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 624.469348] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] return self._sync_wrapper(fn, *args, **kwargs) [ 624.469348] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 624.469348] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] self.wait() [ 624.469348] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 624.469348] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] self[:] = self._gt.wait() [ 624.469348] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 624.469348] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] return self._exit_event.wait() [ 624.469348] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 624.469860] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] result = hub.switch() [ 624.469860] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 624.469860] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] return self.greenlet.switch() [ 624.469860] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 624.469860] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] result = function(*args, **kwargs) [ 624.469860] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 624.469860] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] return func(*args, **kwargs) [ 624.469860] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 624.469860] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] raise e [ 624.469860] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.469860] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] nwinfo = self.network_api.allocate_for_instance( [ 624.469860] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 624.469860] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] created_port_ids = self._update_ports_for_instance( [ 624.470278] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 624.470278] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] with excutils.save_and_reraise_exception(): [ 624.470278] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.470278] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] self.force_reraise() [ 624.470278] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.470278] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] raise self.value [ 624.470278] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 624.470278] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] updated_port = self._update_port( [ 624.470278] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.470278] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] _ensure_no_port_binding_failure(port) [ 624.470278] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.470278] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] raise exception.PortBindingFailed(port_id=port['id']) [ 624.470745] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] nova.exception.PortBindingFailed: Binding failed for port baf9e6b1-3e80-4e7f-9ed6-e6cdabd7320a, please check neutron logs for more information. [ 624.470745] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] [ 624.470745] env[62385]: INFO nova.compute.manager [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Terminating instance [ 624.470877] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] Acquiring lock "refresh_cache-cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.470981] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] Acquired lock "refresh_cache-cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.471111] env[62385]: DEBUG nova.network.neutron [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 624.517736] env[62385]: DEBUG nova.compute.manager [req-8cce8042-ed98-4ec7-b403-e8916933d23c req-08571786-5f97-4e54-bdbe-4aa4dde84fae service nova] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Received event network-changed-baf9e6b1-3e80-4e7f-9ed6-e6cdabd7320a {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 624.518471] env[62385]: DEBUG nova.compute.manager [req-8cce8042-ed98-4ec7-b403-e8916933d23c req-08571786-5f97-4e54-bdbe-4aa4dde84fae service nova] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Refreshing instance network info cache due to event network-changed-baf9e6b1-3e80-4e7f-9ed6-e6cdabd7320a. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 624.518906] env[62385]: DEBUG oslo_concurrency.lockutils [req-8cce8042-ed98-4ec7-b403-e8916933d23c req-08571786-5f97-4e54-bdbe-4aa4dde84fae service nova] Acquiring lock "refresh_cache-cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.699635] env[62385]: DEBUG nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 624.705925] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 624.706237] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ffe0314f-053b-4b5d-803c-bfb53ad167dd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.713937] env[62385]: DEBUG oslo_vmware.api [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Waiting for the task: (returnval){ [ 624.713937] env[62385]: value = "task-1205466" [ 624.713937] env[62385]: _type = "Task" [ 624.713937] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.719109] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.625s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 624.719698] env[62385]: DEBUG nova.compute.manager [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 624.722157] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3d83202f-9f92-4596-a521-3507c75c3596 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.178s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.722388] env[62385]: DEBUG nova.objects.instance [None req-3d83202f-9f92-4596-a521-3507c75c3596 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Lazy-loading 'resources' on Instance uuid e514a72d-6b4d-46f9-935b-3535bed4afc2 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 624.728619] env[62385]: DEBUG oslo_vmware.api [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205466, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.990042] env[62385]: DEBUG nova.network.neutron [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 625.099132] env[62385]: DEBUG nova.network.neutron [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.218548] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.225152] env[62385]: DEBUG nova.compute.utils [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 625.226407] env[62385]: DEBUG oslo_vmware.api [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205466, 'name': PowerOffVM_Task, 'duration_secs': 0.24099} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.226860] env[62385]: DEBUG nova.compute.manager [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 625.227053] env[62385]: DEBUG nova.network.neutron [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 625.228706] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 625.228937] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 625.231703] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6803e9a-3a5c-4e3a-9084-b44d069d1998 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.238638] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 625.238864] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-678c7c45-e2f3-4a5d-a020-91b7214e6bd1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.265897] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 625.266084] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 625.266270] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Deleting the datastore file [datastore1] f0f0eed0-a76a-42bc-9cca-91b93d7ad64c {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 625.266515] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4543d809-3102-46d3-95ba-47d77589b021 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.273778] env[62385]: DEBUG oslo_vmware.api [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Waiting for the task: (returnval){ [ 625.273778] env[62385]: value = "task-1205468" [ 625.273778] env[62385]: _type = "Task" [ 625.273778] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.281901] env[62385]: DEBUG nova.policy [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2f2a72e69a7749cfb5fe080e171cee47', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f7c6fcd602be4c998df22bca228c919d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 625.286267] env[62385]: DEBUG oslo_vmware.api [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205468, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.563202] env[62385]: DEBUG oslo_concurrency.lockutils [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "42ecf594-dff7-4af3-ac56-5bbc45b7a192" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.564069] env[62385]: DEBUG oslo_concurrency.lockutils [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "42ecf594-dff7-4af3-ac56-5bbc45b7a192" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.601553] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] Releasing lock "refresh_cache-cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 625.601968] env[62385]: DEBUG nova.compute.manager [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 625.602237] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 625.602484] env[62385]: DEBUG oslo_concurrency.lockutils [req-8cce8042-ed98-4ec7-b403-e8916933d23c req-08571786-5f97-4e54-bdbe-4aa4dde84fae service nova] Acquired lock "refresh_cache-cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.602653] env[62385]: DEBUG nova.network.neutron [req-8cce8042-ed98-4ec7-b403-e8916933d23c req-08571786-5f97-4e54-bdbe-4aa4dde84fae service nova] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Refreshing network info cache for port baf9e6b1-3e80-4e7f-9ed6-e6cdabd7320a {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 625.603694] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c3b620ac-773a-41a9-9d46-31f1558a35f2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.615115] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bb82470-a402-473e-9769-b4881bab6adf {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.640277] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e could not be found. [ 625.640506] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 625.640686] env[62385]: INFO nova.compute.manager [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 625.640924] env[62385]: DEBUG oslo.service.loopingcall [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 625.643872] env[62385]: DEBUG nova.compute.manager [-] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 625.643964] env[62385]: DEBUG nova.network.neutron [-] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 625.663709] env[62385]: DEBUG nova.network.neutron [-] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 625.713032] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eba596e-6b3e-4c26-a88c-2310f0cc546c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.719292] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b36f0287-0a3d-4abf-baf1-0b66b2a9a815 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.723585] env[62385]: DEBUG nova.network.neutron [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Successfully created port: f51c0b92-691e-4172-8fa3-d6d28cc46bcf {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 625.730027] env[62385]: DEBUG nova.compute.manager [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 625.754773] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aec4a8e-abbc-438a-bac3-cca1a8d8fc02 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.763282] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6956b4c-16dc-4409-a664-6197c15c8552 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.779282] env[62385]: DEBUG nova.compute.provider_tree [None req-3d83202f-9f92-4596-a521-3507c75c3596 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 625.787945] env[62385]: DEBUG oslo_vmware.api [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205468, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.099075} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.788695] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 625.788888] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 625.789075] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 626.124985] env[62385]: DEBUG nova.network.neutron [req-8cce8042-ed98-4ec7-b403-e8916933d23c req-08571786-5f97-4e54-bdbe-4aa4dde84fae service nova] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 626.168399] env[62385]: DEBUG nova.network.neutron [-] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.205534] env[62385]: DEBUG nova.network.neutron [req-8cce8042-ed98-4ec7-b403-e8916933d23c req-08571786-5f97-4e54-bdbe-4aa4dde84fae service nova] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.284549] env[62385]: DEBUG nova.scheduler.client.report [None req-3d83202f-9f92-4596-a521-3507c75c3596 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 626.574023] env[62385]: DEBUG nova.compute.manager [req-c01bf8ef-28bf-461c-aea1-d4d9a62509f9 req-1d5d8243-c183-45d5-a9d2-630979d4b772 service nova] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Received event network-vif-deleted-baf9e6b1-3e80-4e7f-9ed6-e6cdabd7320a {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 626.673217] env[62385]: INFO nova.compute.manager [-] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Took 1.03 seconds to deallocate network for instance. [ 626.677035] env[62385]: DEBUG nova.compute.claims [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Aborting claim: {{(pid=62385) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 626.677035] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.708546] env[62385]: DEBUG oslo_concurrency.lockutils [req-8cce8042-ed98-4ec7-b403-e8916933d23c req-08571786-5f97-4e54-bdbe-4aa4dde84fae service nova] Releasing lock "refresh_cache-cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.760039] env[62385]: DEBUG nova.compute.manager [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 626.783578] env[62385]: DEBUG nova.virt.hardware [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 626.783813] env[62385]: DEBUG nova.virt.hardware [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 626.783966] env[62385]: DEBUG nova.virt.hardware [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 626.784164] env[62385]: DEBUG nova.virt.hardware [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 626.784310] env[62385]: DEBUG nova.virt.hardware [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 626.784461] env[62385]: DEBUG nova.virt.hardware [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 626.784692] env[62385]: DEBUG nova.virt.hardware [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 626.784851] env[62385]: DEBUG nova.virt.hardware [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 626.785036] env[62385]: DEBUG nova.virt.hardware [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 626.785229] env[62385]: DEBUG nova.virt.hardware [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 626.785373] env[62385]: DEBUG nova.virt.hardware [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 626.786261] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4198c1c-fdd9-4a06-ab64-fa6efa73b71d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.789836] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3d83202f-9f92-4596-a521-3507c75c3596 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.068s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 626.792462] env[62385]: DEBUG oslo_concurrency.lockutils [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.500s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 626.801316] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3ebe862-a656-4d98-9c1b-d00f62d3864a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.827671] env[62385]: INFO nova.scheduler.client.report [None req-3d83202f-9f92-4596-a521-3507c75c3596 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Deleted allocations for instance e514a72d-6b4d-46f9-935b-3535bed4afc2 [ 626.842471] env[62385]: DEBUG nova.virt.hardware [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 626.842697] env[62385]: DEBUG nova.virt.hardware [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 626.842853] env[62385]: DEBUG nova.virt.hardware [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 626.843377] env[62385]: DEBUG nova.virt.hardware [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 626.843377] env[62385]: DEBUG nova.virt.hardware [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 626.843377] env[62385]: DEBUG nova.virt.hardware [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 626.843516] env[62385]: DEBUG nova.virt.hardware [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 626.843668] env[62385]: DEBUG nova.virt.hardware [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 626.843839] env[62385]: DEBUG nova.virt.hardware [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 626.843997] env[62385]: DEBUG nova.virt.hardware [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 626.844232] env[62385]: DEBUG nova.virt.hardware [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 626.846695] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8776cd7-355e-4cbb-895c-c0595457596e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.855482] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b77d676b-d406-464d-b2be-08a20911b2ab {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.872142] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Instance VIF info [] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 626.877963] env[62385]: DEBUG oslo.service.loopingcall [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 626.878229] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 626.878991] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-24da63f1-35c7-44bc-b090-730807a9d3b5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.895745] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 626.895745] env[62385]: value = "task-1205469" [ 626.895745] env[62385]: _type = "Task" [ 626.895745] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.904048] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205469, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.055137] env[62385]: ERROR nova.compute.manager [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f51c0b92-691e-4172-8fa3-d6d28cc46bcf, please check neutron logs for more information. [ 627.055137] env[62385]: ERROR nova.compute.manager Traceback (most recent call last): [ 627.055137] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 627.055137] env[62385]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 627.055137] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 627.055137] env[62385]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 627.055137] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 627.055137] env[62385]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 627.055137] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.055137] env[62385]: ERROR nova.compute.manager self.force_reraise() [ 627.055137] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.055137] env[62385]: ERROR nova.compute.manager raise self.value [ 627.055137] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 627.055137] env[62385]: ERROR nova.compute.manager updated_port = self._update_port( [ 627.055137] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.055137] env[62385]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 627.055536] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.055536] env[62385]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 627.055536] env[62385]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f51c0b92-691e-4172-8fa3-d6d28cc46bcf, please check neutron logs for more information. [ 627.055536] env[62385]: ERROR nova.compute.manager [ 627.055536] env[62385]: Traceback (most recent call last): [ 627.055536] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 627.055536] env[62385]: listener.cb(fileno) [ 627.055536] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 627.055536] env[62385]: result = function(*args, **kwargs) [ 627.055536] env[62385]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 627.055536] env[62385]: return func(*args, **kwargs) [ 627.055536] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 627.055536] env[62385]: raise e [ 627.055536] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 627.055536] env[62385]: nwinfo = self.network_api.allocate_for_instance( [ 627.055536] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 627.055536] env[62385]: created_port_ids = self._update_ports_for_instance( [ 627.055536] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 627.055536] env[62385]: with excutils.save_and_reraise_exception(): [ 627.055536] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.055536] env[62385]: self.force_reraise() [ 627.055536] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.055536] env[62385]: raise self.value [ 627.055536] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 627.055536] env[62385]: updated_port = self._update_port( [ 627.055536] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.055536] env[62385]: _ensure_no_port_binding_failure(port) [ 627.055536] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.055536] env[62385]: raise exception.PortBindingFailed(port_id=port['id']) [ 627.056247] env[62385]: nova.exception.PortBindingFailed: Binding failed for port f51c0b92-691e-4172-8fa3-d6d28cc46bcf, please check neutron logs for more information. [ 627.056247] env[62385]: Removing descriptor: 19 [ 627.056462] env[62385]: ERROR nova.compute.manager [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f51c0b92-691e-4172-8fa3-d6d28cc46bcf, please check neutron logs for more information. [ 627.056462] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Traceback (most recent call last): [ 627.056462] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 627.056462] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] yield resources [ 627.056462] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 627.056462] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] self.driver.spawn(context, instance, image_meta, [ 627.056462] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 627.056462] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 627.056462] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 627.056462] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] vm_ref = self.build_virtual_machine(instance, [ 627.056462] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 627.056826] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] vif_infos = vmwarevif.get_vif_info(self._session, [ 627.056826] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 627.056826] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] for vif in network_info: [ 627.056826] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 627.056826] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] return self._sync_wrapper(fn, *args, **kwargs) [ 627.056826] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 627.056826] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] self.wait() [ 627.056826] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 627.056826] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] self[:] = self._gt.wait() [ 627.056826] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 627.056826] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] return self._exit_event.wait() [ 627.056826] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 627.056826] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] result = hub.switch() [ 627.057197] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 627.057197] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] return self.greenlet.switch() [ 627.057197] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 627.057197] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] result = function(*args, **kwargs) [ 627.057197] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 627.057197] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] return func(*args, **kwargs) [ 627.057197] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 627.057197] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] raise e [ 627.057197] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 627.057197] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] nwinfo = self.network_api.allocate_for_instance( [ 627.057197] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 627.057197] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] created_port_ids = self._update_ports_for_instance( [ 627.057197] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 627.057616] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] with excutils.save_and_reraise_exception(): [ 627.057616] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.057616] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] self.force_reraise() [ 627.057616] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.057616] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] raise self.value [ 627.057616] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 627.057616] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] updated_port = self._update_port( [ 627.057616] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.057616] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] _ensure_no_port_binding_failure(port) [ 627.057616] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.057616] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] raise exception.PortBindingFailed(port_id=port['id']) [ 627.057616] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] nova.exception.PortBindingFailed: Binding failed for port f51c0b92-691e-4172-8fa3-d6d28cc46bcf, please check neutron logs for more information. [ 627.057616] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] [ 627.057955] env[62385]: INFO nova.compute.manager [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Terminating instance [ 627.059582] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Acquiring lock "refresh_cache-70abd86c-aa63-4ccb-b185-65bf4977dfdf" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.059797] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Acquired lock "refresh_cache-70abd86c-aa63-4ccb-b185-65bf4977dfdf" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.060045] env[62385]: DEBUG nova.network.neutron [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 627.340669] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3d83202f-9f92-4596-a521-3507c75c3596 tempest-ServerDiagnosticsV248Test-519789768 tempest-ServerDiagnosticsV248Test-519789768-project-member] Lock "e514a72d-6b4d-46f9-935b-3535bed4afc2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.067s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.405643] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205469, 'name': CreateVM_Task, 'duration_secs': 0.253867} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.407843] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 627.408406] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.408565] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.408863] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 627.409140] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-45e01df3-0fb5-4553-a8ae-65e85a9d929f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.413247] env[62385]: DEBUG oslo_vmware.api [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Waiting for the task: (returnval){ [ 627.413247] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e3ba07-7bc0-022c-96bc-dfdd26470cc0" [ 627.413247] env[62385]: _type = "Task" [ 627.413247] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.423100] env[62385]: DEBUG oslo_vmware.api [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e3ba07-7bc0-022c-96bc-dfdd26470cc0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.576431] env[62385]: DEBUG nova.network.neutron [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 627.638639] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bafa8dbd-622e-4462-bf4e-669a53593254 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.643646] env[62385]: DEBUG nova.network.neutron [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.647267] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed8ae4b6-632d-4652-b34f-20f672b9d7b8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.677127] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbe73677-1385-4a24-8d01-0cf692c8d46e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.684464] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1975b54b-9314-4d7f-9c27-4f1a0fe07654 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.699257] env[62385]: DEBUG nova.compute.provider_tree [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 627.923939] env[62385]: DEBUG oslo_vmware.api [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e3ba07-7bc0-022c-96bc-dfdd26470cc0, 'name': SearchDatastore_Task, 'duration_secs': 0.008756} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.924329] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 627.924599] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 627.924890] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.925072] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.925261] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 627.925524] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d5fa2105-14e6-40ac-b43e-b0f0e7437b1c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.933538] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 627.933716] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 627.934645] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-507374c2-170f-478c-98ce-c5a1f3ee1f3d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.939580] env[62385]: DEBUG oslo_vmware.api [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Waiting for the task: (returnval){ [ 627.939580] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52520e16-2d93-fdc3-389b-5aa0821f9ac5" [ 627.939580] env[62385]: _type = "Task" [ 627.939580] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.947800] env[62385]: DEBUG oslo_vmware.api [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52520e16-2d93-fdc3-389b-5aa0821f9ac5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.151634] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Releasing lock "refresh_cache-70abd86c-aa63-4ccb-b185-65bf4977dfdf" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 628.152014] env[62385]: DEBUG nova.compute.manager [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 628.152518] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 628.152518] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7ba8c457-cb94-4e6e-99c4-e6b9dab77326 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.161557] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aff60317-c666-41ea-af81-adcd0da4a278 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.184061] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 70abd86c-aa63-4ccb-b185-65bf4977dfdf could not be found. [ 628.184298] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 628.184485] env[62385]: INFO nova.compute.manager [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Took 0.03 seconds to destroy the instance on the hypervisor. [ 628.184764] env[62385]: DEBUG oslo.service.loopingcall [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 628.184983] env[62385]: DEBUG nova.compute.manager [-] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 628.185091] env[62385]: DEBUG nova.network.neutron [-] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 628.202729] env[62385]: DEBUG nova.scheduler.client.report [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 628.206948] env[62385]: DEBUG nova.network.neutron [-] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 628.451418] env[62385]: DEBUG oslo_vmware.api [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52520e16-2d93-fdc3-389b-5aa0821f9ac5, 'name': SearchDatastore_Task, 'duration_secs': 0.008} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.452216] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-87ecfdc0-7487-4b86-b01e-85dec0757596 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.457609] env[62385]: DEBUG oslo_vmware.api [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Waiting for the task: (returnval){ [ 628.457609] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52857620-3e11-e068-34a1-ab46df04bce2" [ 628.457609] env[62385]: _type = "Task" [ 628.457609] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.465162] env[62385]: DEBUG oslo_vmware.api [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52857620-3e11-e068-34a1-ab46df04bce2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.605491] env[62385]: DEBUG nova.compute.manager [req-776049c1-52e6-4896-8ad5-1943ab75a52d req-b00c4fcf-2e1d-48cd-a9d0-fdadce914d66 service nova] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Received event network-changed-f51c0b92-691e-4172-8fa3-d6d28cc46bcf {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 628.605674] env[62385]: DEBUG nova.compute.manager [req-776049c1-52e6-4896-8ad5-1943ab75a52d req-b00c4fcf-2e1d-48cd-a9d0-fdadce914d66 service nova] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Refreshing instance network info cache due to event network-changed-f51c0b92-691e-4172-8fa3-d6d28cc46bcf. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 628.605879] env[62385]: DEBUG oslo_concurrency.lockutils [req-776049c1-52e6-4896-8ad5-1943ab75a52d req-b00c4fcf-2e1d-48cd-a9d0-fdadce914d66 service nova] Acquiring lock "refresh_cache-70abd86c-aa63-4ccb-b185-65bf4977dfdf" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 628.606023] env[62385]: DEBUG oslo_concurrency.lockutils [req-776049c1-52e6-4896-8ad5-1943ab75a52d req-b00c4fcf-2e1d-48cd-a9d0-fdadce914d66 service nova] Acquired lock "refresh_cache-70abd86c-aa63-4ccb-b185-65bf4977dfdf" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.606182] env[62385]: DEBUG nova.network.neutron [req-776049c1-52e6-4896-8ad5-1943ab75a52d req-b00c4fcf-2e1d-48cd-a9d0-fdadce914d66 service nova] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Refreshing network info cache for port f51c0b92-691e-4172-8fa3-d6d28cc46bcf {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 628.708764] env[62385]: DEBUG oslo_concurrency.lockutils [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.917s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 628.709421] env[62385]: ERROR nova.compute.manager [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c7bdc405-0f30-4e93-a8e5-daffaabdfaf8, please check neutron logs for more information. [ 628.709421] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Traceback (most recent call last): [ 628.709421] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 628.709421] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] self.driver.spawn(context, instance, image_meta, [ 628.709421] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 628.709421] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 628.709421] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 628.709421] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] vm_ref = self.build_virtual_machine(instance, [ 628.709421] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 628.709421] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] vif_infos = vmwarevif.get_vif_info(self._session, [ 628.709421] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 628.709744] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] for vif in network_info: [ 628.709744] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 628.709744] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] return self._sync_wrapper(fn, *args, **kwargs) [ 628.709744] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 628.709744] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] self.wait() [ 628.709744] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 628.709744] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] self[:] = self._gt.wait() [ 628.709744] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 628.709744] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] return self._exit_event.wait() [ 628.709744] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 628.709744] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] result = hub.switch() [ 628.709744] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 628.709744] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] return self.greenlet.switch() [ 628.710116] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 628.710116] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] result = function(*args, **kwargs) [ 628.710116] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 628.710116] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] return func(*args, **kwargs) [ 628.710116] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 628.710116] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] raise e [ 628.710116] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 628.710116] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] nwinfo = self.network_api.allocate_for_instance( [ 628.710116] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 628.710116] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] created_port_ids = self._update_ports_for_instance( [ 628.710116] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 628.710116] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] with excutils.save_and_reraise_exception(): [ 628.710116] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 628.710567] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] self.force_reraise() [ 628.710567] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 628.710567] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] raise self.value [ 628.710567] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 628.710567] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] updated_port = self._update_port( [ 628.710567] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 628.710567] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] _ensure_no_port_binding_failure(port) [ 628.710567] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 628.710567] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] raise exception.PortBindingFailed(port_id=port['id']) [ 628.710567] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] nova.exception.PortBindingFailed: Binding failed for port c7bdc405-0f30-4e93-a8e5-daffaabdfaf8, please check neutron logs for more information. [ 628.710567] env[62385]: ERROR nova.compute.manager [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] [ 628.710943] env[62385]: DEBUG nova.compute.utils [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Binding failed for port c7bdc405-0f30-4e93-a8e5-daffaabdfaf8, please check neutron logs for more information. {{(pid=62385) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 628.711478] env[62385]: DEBUG nova.network.neutron [-] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.712531] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.004s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 628.714353] env[62385]: INFO nova.compute.claims [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 628.718063] env[62385]: DEBUG nova.compute.manager [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Build of instance 4a99d2e8-7795-4016-ad6e-39a816c7fae1 was re-scheduled: Binding failed for port c7bdc405-0f30-4e93-a8e5-daffaabdfaf8, please check neutron logs for more information. {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 628.718528] env[62385]: DEBUG nova.compute.manager [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Unplugging VIFs for instance {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 628.719538] env[62385]: DEBUG oslo_concurrency.lockutils [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Acquiring lock "refresh_cache-4a99d2e8-7795-4016-ad6e-39a816c7fae1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 628.719538] env[62385]: DEBUG oslo_concurrency.lockutils [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Acquired lock "refresh_cache-4a99d2e8-7795-4016-ad6e-39a816c7fae1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.719538] env[62385]: DEBUG nova.network.neutron [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 628.969290] env[62385]: DEBUG oslo_vmware.api [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52857620-3e11-e068-34a1-ab46df04bce2, 'name': SearchDatastore_Task, 'duration_secs': 0.007708} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.969547] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 628.969813] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] f0f0eed0-a76a-42bc-9cca-91b93d7ad64c/f0f0eed0-a76a-42bc-9cca-91b93d7ad64c.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 628.970062] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d9aafedc-0ff4-44f8-9402-41b7bed5ecc1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.976413] env[62385]: DEBUG oslo_vmware.api [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Waiting for the task: (returnval){ [ 628.976413] env[62385]: value = "task-1205470" [ 628.976413] env[62385]: _type = "Task" [ 628.976413] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.984709] env[62385]: DEBUG oslo_vmware.api [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205470, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.125450] env[62385]: DEBUG nova.network.neutron [req-776049c1-52e6-4896-8ad5-1943ab75a52d req-b00c4fcf-2e1d-48cd-a9d0-fdadce914d66 service nova] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 629.187448] env[62385]: DEBUG nova.network.neutron [req-776049c1-52e6-4896-8ad5-1943ab75a52d req-b00c4fcf-2e1d-48cd-a9d0-fdadce914d66 service nova] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.221677] env[62385]: INFO nova.compute.manager [-] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Took 1.04 seconds to deallocate network for instance. [ 629.228384] env[62385]: DEBUG nova.compute.claims [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Aborting claim: {{(pid=62385) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 629.229034] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.253762] env[62385]: DEBUG nova.network.neutron [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 629.405381] env[62385]: DEBUG nova.network.neutron [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.487492] env[62385]: DEBUG oslo_vmware.api [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205470, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.45308} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.488778] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] f0f0eed0-a76a-42bc-9cca-91b93d7ad64c/f0f0eed0-a76a-42bc-9cca-91b93d7ad64c.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 629.488778] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 629.488778] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-43074b2a-f8b5-4f74-b3d0-a1a22db291ed {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.498021] env[62385]: DEBUG oslo_vmware.api [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Waiting for the task: (returnval){ [ 629.498021] env[62385]: value = "task-1205471" [ 629.498021] env[62385]: _type = "Task" [ 629.498021] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.507167] env[62385]: DEBUG oslo_vmware.api [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205471, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.689885] env[62385]: DEBUG oslo_concurrency.lockutils [req-776049c1-52e6-4896-8ad5-1943ab75a52d req-b00c4fcf-2e1d-48cd-a9d0-fdadce914d66 service nova] Releasing lock "refresh_cache-70abd86c-aa63-4ccb-b185-65bf4977dfdf" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.691028] env[62385]: DEBUG nova.compute.manager [req-776049c1-52e6-4896-8ad5-1943ab75a52d req-b00c4fcf-2e1d-48cd-a9d0-fdadce914d66 service nova] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Received event network-vif-deleted-f51c0b92-691e-4172-8fa3-d6d28cc46bcf {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 629.907589] env[62385]: DEBUG oslo_concurrency.lockutils [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Releasing lock "refresh_cache-4a99d2e8-7795-4016-ad6e-39a816c7fae1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.907837] env[62385]: DEBUG nova.compute.manager [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 629.908023] env[62385]: DEBUG nova.compute.manager [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 629.908196] env[62385]: DEBUG nova.network.neutron [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 629.924323] env[62385]: DEBUG nova.network.neutron [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 630.007219] env[62385]: DEBUG oslo_vmware.api [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205471, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.211898} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.007490] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 630.011212] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d0d7bb6-cfda-4d36-b17d-0403040124a8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.031529] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Reconfiguring VM instance instance-00000019 to attach disk [datastore1] f0f0eed0-a76a-42bc-9cca-91b93d7ad64c/f0f0eed0-a76a-42bc-9cca-91b93d7ad64c.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 630.032140] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fa177bda-8f3d-4386-b12c-40aa5f81eb06 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.058018] env[62385]: DEBUG oslo_vmware.api [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Waiting for the task: (returnval){ [ 630.058018] env[62385]: value = "task-1205472" [ 630.058018] env[62385]: _type = "Task" [ 630.058018] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.066719] env[62385]: DEBUG oslo_vmware.api [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205472, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.170017] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66c585fe-bc1e-45cb-9eef-987e154256c8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.175701] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90230d2d-00db-46fb-9520-07f127a203ac {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.206618] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b330689a-197a-4ef2-b893-bf4d0f8d1109 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.213890] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f54baa58-4355-499c-b7e8-509f00ab2c40 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.227071] env[62385]: DEBUG nova.compute.provider_tree [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 630.430631] env[62385]: DEBUG nova.network.neutron [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.566751] env[62385]: DEBUG oslo_vmware.api [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205472, 'name': ReconfigVM_Task, 'duration_secs': 0.256906} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.566989] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Reconfigured VM instance instance-00000019 to attach disk [datastore1] f0f0eed0-a76a-42bc-9cca-91b93d7ad64c/f0f0eed0-a76a-42bc-9cca-91b93d7ad64c.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 630.567707] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aa07bfbb-89fb-48a4-89f7-5b56ffb2fcef {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.574953] env[62385]: DEBUG oslo_vmware.api [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Waiting for the task: (returnval){ [ 630.574953] env[62385]: value = "task-1205473" [ 630.574953] env[62385]: _type = "Task" [ 630.574953] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.582978] env[62385]: DEBUG oslo_vmware.api [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205473, 'name': Rename_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.730880] env[62385]: DEBUG nova.scheduler.client.report [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 630.934710] env[62385]: INFO nova.compute.manager [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: 4a99d2e8-7795-4016-ad6e-39a816c7fae1] Took 1.03 seconds to deallocate network for instance. [ 631.087611] env[62385]: DEBUG oslo_vmware.api [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205473, 'name': Rename_Task, 'duration_secs': 0.118553} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.087611] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 631.087867] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-66b36a8e-c42d-4bc2-80f4-5cd9b964073c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.094750] env[62385]: DEBUG oslo_vmware.api [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Waiting for the task: (returnval){ [ 631.094750] env[62385]: value = "task-1205474" [ 631.094750] env[62385]: _type = "Task" [ 631.094750] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.113098] env[62385]: DEBUG oslo_vmware.api [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205474, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.236427] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.523s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.236427] env[62385]: DEBUG nova.compute.manager [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 631.242885] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.066s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.604590] env[62385]: DEBUG oslo_vmware.api [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205474, 'name': PowerOnVM_Task, 'duration_secs': 0.408681} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.604929] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 631.605152] env[62385]: DEBUG nova.compute.manager [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 631.605926] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5028633d-d8b5-40f7-bc91-097964648f39 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.748308] env[62385]: DEBUG nova.compute.utils [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 631.749709] env[62385]: DEBUG nova.compute.manager [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 631.749874] env[62385]: DEBUG nova.network.neutron [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 631.821250] env[62385]: DEBUG nova.policy [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bc2a16eb9c544ccdb7df43bdd542e416', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0730f863135b4f819b381e385dc30e56', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 631.979282] env[62385]: INFO nova.scheduler.client.report [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Deleted allocations for instance 4a99d2e8-7795-4016-ad6e-39a816c7fae1 [ 632.126058] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.255340] env[62385]: DEBUG nova.compute.manager [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 632.318154] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5523755a-f2a3-402b-8535-0d096ed99ea0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.321925] env[62385]: DEBUG nova.network.neutron [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Successfully created port: 99d91aa3-3018-4ac1-b1c4-c3f41b5cf2f9 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 632.328876] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1cef222-e7f4-416a-b010-726f65f87249 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.362836] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b30ac80-0753-430b-ac83-ddc36b863304 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.371931] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0788e969-2ce3-433b-964a-4a6b81f4f731 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.387932] env[62385]: DEBUG nova.compute.provider_tree [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 632.490583] env[62385]: DEBUG oslo_concurrency.lockutils [None req-641e8bfd-8599-4c56-b0e1-77df7bff908c tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Lock "4a99d2e8-7795-4016-ad6e-39a816c7fae1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 93.737s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 632.868543] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Acquiring lock "a5c4afc6-38a6-4815-8ec4-cc01c24489bf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.868795] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Lock "a5c4afc6-38a6-4815-8ec4-cc01c24489bf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 632.890614] env[62385]: DEBUG nova.scheduler.client.report [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 632.994023] env[62385]: DEBUG nova.compute.manager [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 633.272531] env[62385]: DEBUG nova.compute.manager [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 633.308033] env[62385]: DEBUG nova.virt.hardware [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 633.308033] env[62385]: DEBUG nova.virt.hardware [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 633.308033] env[62385]: DEBUG nova.virt.hardware [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 633.308783] env[62385]: DEBUG nova.virt.hardware [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 633.308783] env[62385]: DEBUG nova.virt.hardware [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 633.308783] env[62385]: DEBUG nova.virt.hardware [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 633.308783] env[62385]: DEBUG nova.virt.hardware [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 633.309227] env[62385]: DEBUG nova.virt.hardware [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 633.309772] env[62385]: DEBUG nova.virt.hardware [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 633.310080] env[62385]: DEBUG nova.virt.hardware [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 633.310410] env[62385]: DEBUG nova.virt.hardware [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 633.313100] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d54b33e-f7b5-4fae-8c39-18ef82226cff {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.320912] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79ca2e90-e0cd-433f-b8cf-a6d049b38952 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.341254] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ae8492d9-8f89-48fe-aa8c-4b169b2b533c tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Acquiring lock "f0f0eed0-a76a-42bc-9cca-91b93d7ad64c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.341766] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ae8492d9-8f89-48fe-aa8c-4b169b2b533c tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Lock "f0f0eed0-a76a-42bc-9cca-91b93d7ad64c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.342131] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ae8492d9-8f89-48fe-aa8c-4b169b2b533c tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Acquiring lock "f0f0eed0-a76a-42bc-9cca-91b93d7ad64c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.342451] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ae8492d9-8f89-48fe-aa8c-4b169b2b533c tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Lock "f0f0eed0-a76a-42bc-9cca-91b93d7ad64c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.342743] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ae8492d9-8f89-48fe-aa8c-4b169b2b533c tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Lock "f0f0eed0-a76a-42bc-9cca-91b93d7ad64c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 633.345293] env[62385]: INFO nova.compute.manager [None req-ae8492d9-8f89-48fe-aa8c-4b169b2b533c tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Terminating instance [ 633.348620] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ae8492d9-8f89-48fe-aa8c-4b169b2b533c tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Acquiring lock "refresh_cache-f0f0eed0-a76a-42bc-9cca-91b93d7ad64c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 633.348620] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ae8492d9-8f89-48fe-aa8c-4b169b2b533c tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Acquired lock "refresh_cache-f0f0eed0-a76a-42bc-9cca-91b93d7ad64c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.348620] env[62385]: DEBUG nova.network.neutron [None req-ae8492d9-8f89-48fe-aa8c-4b169b2b533c tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 633.396900] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.154s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 633.397185] env[62385]: ERROR nova.compute.manager [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1fac3d82-702d-4ce0-906e-0e251b78b679, please check neutron logs for more information. [ 633.397185] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Traceback (most recent call last): [ 633.397185] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 633.397185] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] self.driver.spawn(context, instance, image_meta, [ 633.397185] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 633.397185] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] self._vmops.spawn(context, instance, image_meta, injected_files, [ 633.397185] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 633.397185] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] vm_ref = self.build_virtual_machine(instance, [ 633.397185] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 633.397185] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] vif_infos = vmwarevif.get_vif_info(self._session, [ 633.397185] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 633.397522] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] for vif in network_info: [ 633.397522] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 633.397522] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] return self._sync_wrapper(fn, *args, **kwargs) [ 633.397522] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 633.397522] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] self.wait() [ 633.397522] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 633.397522] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] self[:] = self._gt.wait() [ 633.397522] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 633.397522] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] return self._exit_event.wait() [ 633.397522] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 633.397522] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] current.throw(*self._exc) [ 633.397522] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 633.397522] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] result = function(*args, **kwargs) [ 633.397915] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 633.397915] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] return func(*args, **kwargs) [ 633.397915] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 633.397915] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] raise e [ 633.397915] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 633.397915] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] nwinfo = self.network_api.allocate_for_instance( [ 633.397915] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 633.397915] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] created_port_ids = self._update_ports_for_instance( [ 633.397915] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 633.397915] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] with excutils.save_and_reraise_exception(): [ 633.397915] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.397915] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] self.force_reraise() [ 633.397915] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.398267] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] raise self.value [ 633.398267] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 633.398267] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] updated_port = self._update_port( [ 633.398267] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 633.398267] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] _ensure_no_port_binding_failure(port) [ 633.398267] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 633.398267] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] raise exception.PortBindingFailed(port_id=port['id']) [ 633.398267] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] nova.exception.PortBindingFailed: Binding failed for port 1fac3d82-702d-4ce0-906e-0e251b78b679, please check neutron logs for more information. [ 633.398267] env[62385]: ERROR nova.compute.manager [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] [ 633.399548] env[62385]: DEBUG nova.compute.utils [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Binding failed for port 1fac3d82-702d-4ce0-906e-0e251b78b679, please check neutron logs for more information. {{(pid=62385) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 633.403027] env[62385]: DEBUG oslo_concurrency.lockutils [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.407s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.403027] env[62385]: INFO nova.compute.claims [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] [instance: ad12e5b0-6854-4281-9155-e1660393829c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 633.409710] env[62385]: DEBUG nova.compute.manager [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Build of instance 317e7e7c-7c6f-4855-a433-e55ea5571b92 was re-scheduled: Binding failed for port 1fac3d82-702d-4ce0-906e-0e251b78b679, please check neutron logs for more information. {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 633.409902] env[62385]: DEBUG nova.compute.manager [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Unplugging VIFs for instance {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 633.410299] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] Acquiring lock "refresh_cache-317e7e7c-7c6f-4855-a433-e55ea5571b92" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 633.410468] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] Acquired lock "refresh_cache-317e7e7c-7c6f-4855-a433-e55ea5571b92" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.410632] env[62385]: DEBUG nova.network.neutron [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 633.512717] env[62385]: DEBUG nova.compute.manager [req-9c472c4f-efe4-4536-a675-46dfa8148d8d req-b077e36e-062f-4975-a704-e62b52bbfca7 service nova] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Received event network-changed-99d91aa3-3018-4ac1-b1c4-c3f41b5cf2f9 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 633.512963] env[62385]: DEBUG nova.compute.manager [req-9c472c4f-efe4-4536-a675-46dfa8148d8d req-b077e36e-062f-4975-a704-e62b52bbfca7 service nova] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Refreshing instance network info cache due to event network-changed-99d91aa3-3018-4ac1-b1c4-c3f41b5cf2f9. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 633.513252] env[62385]: DEBUG oslo_concurrency.lockutils [req-9c472c4f-efe4-4536-a675-46dfa8148d8d req-b077e36e-062f-4975-a704-e62b52bbfca7 service nova] Acquiring lock "refresh_cache-c348ef22-abb0-48cf-98a6-3a4b217dd186" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 633.513436] env[62385]: DEBUG oslo_concurrency.lockutils [req-9c472c4f-efe4-4536-a675-46dfa8148d8d req-b077e36e-062f-4975-a704-e62b52bbfca7 service nova] Acquired lock "refresh_cache-c348ef22-abb0-48cf-98a6-3a4b217dd186" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.513632] env[62385]: DEBUG nova.network.neutron [req-9c472c4f-efe4-4536-a675-46dfa8148d8d req-b077e36e-062f-4975-a704-e62b52bbfca7 service nova] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Refreshing network info cache for port 99d91aa3-3018-4ac1-b1c4-c3f41b5cf2f9 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 633.523617] env[62385]: DEBUG oslo_concurrency.lockutils [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.750563] env[62385]: ERROR nova.compute.manager [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 99d91aa3-3018-4ac1-b1c4-c3f41b5cf2f9, please check neutron logs for more information. [ 633.750563] env[62385]: ERROR nova.compute.manager Traceback (most recent call last): [ 633.750563] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 633.750563] env[62385]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 633.750563] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 633.750563] env[62385]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 633.750563] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 633.750563] env[62385]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 633.750563] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.750563] env[62385]: ERROR nova.compute.manager self.force_reraise() [ 633.750563] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.750563] env[62385]: ERROR nova.compute.manager raise self.value [ 633.750563] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 633.750563] env[62385]: ERROR nova.compute.manager updated_port = self._update_port( [ 633.750563] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 633.750563] env[62385]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 633.751064] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 633.751064] env[62385]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 633.751064] env[62385]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 99d91aa3-3018-4ac1-b1c4-c3f41b5cf2f9, please check neutron logs for more information. [ 633.751064] env[62385]: ERROR nova.compute.manager [ 633.751064] env[62385]: Traceback (most recent call last): [ 633.751064] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 633.751064] env[62385]: listener.cb(fileno) [ 633.751064] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 633.751064] env[62385]: result = function(*args, **kwargs) [ 633.751064] env[62385]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 633.751064] env[62385]: return func(*args, **kwargs) [ 633.751064] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 633.751064] env[62385]: raise e [ 633.751064] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 633.751064] env[62385]: nwinfo = self.network_api.allocate_for_instance( [ 633.751064] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 633.751064] env[62385]: created_port_ids = self._update_ports_for_instance( [ 633.751064] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 633.751064] env[62385]: with excutils.save_and_reraise_exception(): [ 633.751064] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.751064] env[62385]: self.force_reraise() [ 633.751064] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.751064] env[62385]: raise self.value [ 633.751064] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 633.751064] env[62385]: updated_port = self._update_port( [ 633.751064] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 633.751064] env[62385]: _ensure_no_port_binding_failure(port) [ 633.751064] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 633.751064] env[62385]: raise exception.PortBindingFailed(port_id=port['id']) [ 633.751902] env[62385]: nova.exception.PortBindingFailed: Binding failed for port 99d91aa3-3018-4ac1-b1c4-c3f41b5cf2f9, please check neutron logs for more information. [ 633.751902] env[62385]: Removing descriptor: 19 [ 633.751902] env[62385]: ERROR nova.compute.manager [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 99d91aa3-3018-4ac1-b1c4-c3f41b5cf2f9, please check neutron logs for more information. [ 633.751902] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Traceback (most recent call last): [ 633.751902] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 633.751902] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] yield resources [ 633.751902] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 633.751902] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] self.driver.spawn(context, instance, image_meta, [ 633.751902] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 633.751902] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] self._vmops.spawn(context, instance, image_meta, injected_files, [ 633.751902] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 633.751902] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] vm_ref = self.build_virtual_machine(instance, [ 633.752343] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 633.752343] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] vif_infos = vmwarevif.get_vif_info(self._session, [ 633.752343] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 633.752343] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] for vif in network_info: [ 633.752343] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 633.752343] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] return self._sync_wrapper(fn, *args, **kwargs) [ 633.752343] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 633.752343] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] self.wait() [ 633.752343] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 633.752343] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] self[:] = self._gt.wait() [ 633.752343] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 633.752343] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] return self._exit_event.wait() [ 633.752343] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 633.752707] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] result = hub.switch() [ 633.752707] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 633.752707] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] return self.greenlet.switch() [ 633.752707] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 633.752707] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] result = function(*args, **kwargs) [ 633.752707] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 633.752707] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] return func(*args, **kwargs) [ 633.752707] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 633.752707] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] raise e [ 633.752707] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 633.752707] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] nwinfo = self.network_api.allocate_for_instance( [ 633.752707] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 633.752707] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] created_port_ids = self._update_ports_for_instance( [ 633.753196] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 633.753196] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] with excutils.save_and_reraise_exception(): [ 633.753196] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.753196] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] self.force_reraise() [ 633.753196] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.753196] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] raise self.value [ 633.753196] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 633.753196] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] updated_port = self._update_port( [ 633.753196] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 633.753196] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] _ensure_no_port_binding_failure(port) [ 633.753196] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 633.753196] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] raise exception.PortBindingFailed(port_id=port['id']) [ 633.753544] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] nova.exception.PortBindingFailed: Binding failed for port 99d91aa3-3018-4ac1-b1c4-c3f41b5cf2f9, please check neutron logs for more information. [ 633.753544] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] [ 633.753544] env[62385]: INFO nova.compute.manager [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Terminating instance [ 633.754418] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] Acquiring lock "refresh_cache-c348ef22-abb0-48cf-98a6-3a4b217dd186" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 633.865540] env[62385]: DEBUG nova.network.neutron [None req-ae8492d9-8f89-48fe-aa8c-4b169b2b533c tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 633.937703] env[62385]: DEBUG nova.network.neutron [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 633.956716] env[62385]: DEBUG nova.network.neutron [None req-ae8492d9-8f89-48fe-aa8c-4b169b2b533c tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.030418] env[62385]: DEBUG nova.network.neutron [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.032934] env[62385]: DEBUG nova.network.neutron [req-9c472c4f-efe4-4536-a675-46dfa8148d8d req-b077e36e-062f-4975-a704-e62b52bbfca7 service nova] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 634.162304] env[62385]: DEBUG nova.network.neutron [req-9c472c4f-efe4-4536-a675-46dfa8148d8d req-b077e36e-062f-4975-a704-e62b52bbfca7 service nova] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.459871] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ae8492d9-8f89-48fe-aa8c-4b169b2b533c tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Releasing lock "refresh_cache-f0f0eed0-a76a-42bc-9cca-91b93d7ad64c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 634.460328] env[62385]: DEBUG nova.compute.manager [None req-ae8492d9-8f89-48fe-aa8c-4b169b2b533c tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 634.460481] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ae8492d9-8f89-48fe-aa8c-4b169b2b533c tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 634.463988] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-159ac3cd-83e2-4e7b-976c-f1057dcae584 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.471732] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae8492d9-8f89-48fe-aa8c-4b169b2b533c tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 634.471978] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-700b0cb5-fc93-468e-85ee-0a783899e7f0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.481884] env[62385]: DEBUG oslo_vmware.api [None req-ae8492d9-8f89-48fe-aa8c-4b169b2b533c tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Waiting for the task: (returnval){ [ 634.481884] env[62385]: value = "task-1205475" [ 634.481884] env[62385]: _type = "Task" [ 634.481884] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.488996] env[62385]: DEBUG oslo_vmware.api [None req-ae8492d9-8f89-48fe-aa8c-4b169b2b533c tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205475, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.532906] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] Releasing lock "refresh_cache-317e7e7c-7c6f-4855-a433-e55ea5571b92" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 634.533175] env[62385]: DEBUG nova.compute.manager [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 634.533361] env[62385]: DEBUG nova.compute.manager [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 634.533532] env[62385]: DEBUG nova.network.neutron [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 634.559055] env[62385]: DEBUG nova.network.neutron [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 634.668187] env[62385]: DEBUG oslo_concurrency.lockutils [req-9c472c4f-efe4-4536-a675-46dfa8148d8d req-b077e36e-062f-4975-a704-e62b52bbfca7 service nova] Releasing lock "refresh_cache-c348ef22-abb0-48cf-98a6-3a4b217dd186" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 634.668826] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] Acquired lock "refresh_cache-c348ef22-abb0-48cf-98a6-3a4b217dd186" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.669056] env[62385]: DEBUG nova.network.neutron [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 634.831498] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c925926-27d3-45f6-b62e-a7dfa0ab6d76 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.838849] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e8a0715-e75f-457c-a8d9-6f43eb9ad6dd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.870655] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc849176-48b3-425f-ab03-a4029cd5b4c9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.877769] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4580d259-d38c-437f-9a39-10b22d61afb7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.891367] env[62385]: DEBUG nova.compute.provider_tree [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 634.992928] env[62385]: DEBUG oslo_vmware.api [None req-ae8492d9-8f89-48fe-aa8c-4b169b2b533c tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205475, 'name': PowerOffVM_Task, 'duration_secs': 0.194032} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.993206] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae8492d9-8f89-48fe-aa8c-4b169b2b533c tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 634.994037] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ae8492d9-8f89-48fe-aa8c-4b169b2b533c tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 634.994037] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9b8b4595-5461-4430-b125-336cc8bbf8eb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.018767] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ae8492d9-8f89-48fe-aa8c-4b169b2b533c tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 635.018991] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ae8492d9-8f89-48fe-aa8c-4b169b2b533c tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 635.019188] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae8492d9-8f89-48fe-aa8c-4b169b2b533c tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Deleting the datastore file [datastore1] f0f0eed0-a76a-42bc-9cca-91b93d7ad64c {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 635.019455] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0a29726c-e2c1-4567-9202-d37609e247fb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.025969] env[62385]: DEBUG oslo_vmware.api [None req-ae8492d9-8f89-48fe-aa8c-4b169b2b533c tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Waiting for the task: (returnval){ [ 635.025969] env[62385]: value = "task-1205477" [ 635.025969] env[62385]: _type = "Task" [ 635.025969] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.033911] env[62385]: DEBUG oslo_vmware.api [None req-ae8492d9-8f89-48fe-aa8c-4b169b2b533c tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205477, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.063511] env[62385]: DEBUG nova.network.neutron [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.329997] env[62385]: DEBUG nova.network.neutron [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 635.394772] env[62385]: DEBUG nova.scheduler.client.report [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 635.472288] env[62385]: DEBUG nova.network.neutron [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.535397] env[62385]: DEBUG oslo_vmware.api [None req-ae8492d9-8f89-48fe-aa8c-4b169b2b533c tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205477, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.097682} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.535666] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae8492d9-8f89-48fe-aa8c-4b169b2b533c tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 635.535917] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ae8492d9-8f89-48fe-aa8c-4b169b2b533c tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 635.536082] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ae8492d9-8f89-48fe-aa8c-4b169b2b533c tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 635.536202] env[62385]: INFO nova.compute.manager [None req-ae8492d9-8f89-48fe-aa8c-4b169b2b533c tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Took 1.08 seconds to destroy the instance on the hypervisor. [ 635.536437] env[62385]: DEBUG oslo.service.loopingcall [None req-ae8492d9-8f89-48fe-aa8c-4b169b2b533c tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 635.536624] env[62385]: DEBUG nova.compute.manager [-] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 635.536745] env[62385]: DEBUG nova.network.neutron [-] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 635.549905] env[62385]: DEBUG nova.network.neutron [-] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 635.562044] env[62385]: DEBUG nova.compute.manager [req-4d9dbfa9-1b99-4c5f-bcbd-d9de7c6e6136 req-8830d5e6-3905-4e6b-817c-22c78b5de6af service nova] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Received event network-vif-deleted-99d91aa3-3018-4ac1-b1c4-c3f41b5cf2f9 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 635.565831] env[62385]: INFO nova.compute.manager [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] [instance: 317e7e7c-7c6f-4855-a433-e55ea5571b92] Took 1.03 seconds to deallocate network for instance. [ 635.899617] env[62385]: DEBUG oslo_concurrency.lockutils [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.499s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 635.900155] env[62385]: DEBUG nova.compute.manager [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] [instance: ad12e5b0-6854-4281-9155-e1660393829c] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 635.902618] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 16.347s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 635.981125] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] Releasing lock "refresh_cache-c348ef22-abb0-48cf-98a6-3a4b217dd186" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 635.981125] env[62385]: DEBUG nova.compute.manager [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 635.981125] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 635.981125] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f7415495-9595-4d2a-a9bc-162bd2b90588 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.985030] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d5deb37-906c-4805-b9e0-4c3497f7e121 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.009347] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c348ef22-abb0-48cf-98a6-3a4b217dd186 could not be found. [ 636.009601] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 636.009751] env[62385]: INFO nova.compute.manager [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Took 0.03 seconds to destroy the instance on the hypervisor. [ 636.010038] env[62385]: DEBUG oslo.service.loopingcall [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 636.010223] env[62385]: DEBUG nova.compute.manager [-] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 636.010316] env[62385]: DEBUG nova.network.neutron [-] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 636.028179] env[62385]: DEBUG nova.network.neutron [-] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 636.052710] env[62385]: DEBUG nova.network.neutron [-] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.408501] env[62385]: DEBUG nova.compute.utils [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 636.413404] env[62385]: DEBUG nova.compute.manager [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] [instance: ad12e5b0-6854-4281-9155-e1660393829c] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 636.413845] env[62385]: DEBUG nova.network.neutron [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] [instance: ad12e5b0-6854-4281-9155-e1660393829c] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 636.477506] env[62385]: DEBUG nova.policy [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fcc421bbc5a84d479eaae68f018c2396', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6219b0ba865740d4b1c3743cd8f8d088', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 636.532136] env[62385]: DEBUG nova.network.neutron [-] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.556740] env[62385]: INFO nova.compute.manager [-] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Took 1.02 seconds to deallocate network for instance. [ 636.603708] env[62385]: INFO nova.scheduler.client.report [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] Deleted allocations for instance 317e7e7c-7c6f-4855-a433-e55ea5571b92 [ 636.917405] env[62385]: DEBUG nova.compute.manager [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] [instance: ad12e5b0-6854-4281-9155-e1660393829c] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 636.925107] env[62385]: DEBUG nova.network.neutron [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] [instance: ad12e5b0-6854-4281-9155-e1660393829c] Successfully created port: be90e400-b8be-4151-9773-122efa78f030 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 636.950433] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 0ba32627-c646-477d-9062-b84624d01e48 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 636.951622] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance f0f0eed0-a76a-42bc-9cca-91b93d7ad64c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 636.952016] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 01d010dc-6b5b-470b-a387-f95a17600e19 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 636.952281] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 636.952500] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 70abd86c-aa63-4ccb-b185-65bf4977dfdf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 636.952772] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance c348ef22-abb0-48cf-98a6-3a4b217dd186 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 636.953036] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance ad12e5b0-6854-4281-9155-e1660393829c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 636.992806] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Acquiring lock "e73ecefc-fed3-4ccd-88d3-399a4b72bb4b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 636.993338] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Lock "e73ecefc-fed3-4ccd-88d3-399a4b72bb4b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.033802] env[62385]: INFO nova.compute.manager [-] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Took 1.02 seconds to deallocate network for instance. [ 637.036761] env[62385]: DEBUG nova.compute.claims [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Aborting claim: {{(pid=62385) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 637.036997] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 637.065697] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ae8492d9-8f89-48fe-aa8c-4b169b2b533c tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 637.116481] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f6c554f3-4e31-400d-968c-7f6ee5b744cc tempest-ServersTestBootFromVolume-1117303079 tempest-ServersTestBootFromVolume-1117303079-project-member] Lock "317e7e7c-7c6f-4855-a433-e55ea5571b92" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 98.223s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 637.456898] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance da94a435-b36c-42b2-8c84-564265872048 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 637.619674] env[62385]: DEBUG nova.compute.manager [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 637.933000] env[62385]: DEBUG nova.compute.manager [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] [instance: ad12e5b0-6854-4281-9155-e1660393829c] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 637.963845] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 7b159d98-9fd2-4eca-9755-cd2cac502798 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 637.970296] env[62385]: DEBUG nova.compute.manager [req-4265e5d3-988d-41cb-b55b-5f2fa86e2674 req-1ea7599d-cc03-4013-a9b4-1dcca1816887 service nova] [instance: ad12e5b0-6854-4281-9155-e1660393829c] Received event network-changed-be90e400-b8be-4151-9773-122efa78f030 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 637.970296] env[62385]: DEBUG nova.compute.manager [req-4265e5d3-988d-41cb-b55b-5f2fa86e2674 req-1ea7599d-cc03-4013-a9b4-1dcca1816887 service nova] [instance: ad12e5b0-6854-4281-9155-e1660393829c] Refreshing instance network info cache due to event network-changed-be90e400-b8be-4151-9773-122efa78f030. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 637.970296] env[62385]: DEBUG oslo_concurrency.lockutils [req-4265e5d3-988d-41cb-b55b-5f2fa86e2674 req-1ea7599d-cc03-4013-a9b4-1dcca1816887 service nova] Acquiring lock "refresh_cache-ad12e5b0-6854-4281-9155-e1660393829c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 637.970296] env[62385]: DEBUG oslo_concurrency.lockutils [req-4265e5d3-988d-41cb-b55b-5f2fa86e2674 req-1ea7599d-cc03-4013-a9b4-1dcca1816887 service nova] Acquired lock "refresh_cache-ad12e5b0-6854-4281-9155-e1660393829c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.970296] env[62385]: DEBUG nova.network.neutron [req-4265e5d3-988d-41cb-b55b-5f2fa86e2674 req-1ea7599d-cc03-4013-a9b4-1dcca1816887 service nova] [instance: ad12e5b0-6854-4281-9155-e1660393829c] Refreshing network info cache for port be90e400-b8be-4151-9773-122efa78f030 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 637.975013] env[62385]: DEBUG nova.virt.hardware [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 637.975260] env[62385]: DEBUG nova.virt.hardware [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 637.975463] env[62385]: DEBUG nova.virt.hardware [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 637.975759] env[62385]: DEBUG nova.virt.hardware [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 637.975987] env[62385]: DEBUG nova.virt.hardware [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 637.976819] env[62385]: DEBUG nova.virt.hardware [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 637.976819] env[62385]: DEBUG nova.virt.hardware [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 637.976819] env[62385]: DEBUG nova.virt.hardware [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 637.976819] env[62385]: DEBUG nova.virt.hardware [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 637.977075] env[62385]: DEBUG nova.virt.hardware [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 637.977266] env[62385]: DEBUG nova.virt.hardware [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 637.978941] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6553c86f-d9f4-40be-95a9-5da2966b335f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.990053] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9387efbd-f1c5-4e43-8c25-c99515b2d84d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.170383] env[62385]: DEBUG oslo_concurrency.lockutils [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.187134] env[62385]: ERROR nova.compute.manager [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port be90e400-b8be-4151-9773-122efa78f030, please check neutron logs for more information. [ 638.187134] env[62385]: ERROR nova.compute.manager Traceback (most recent call last): [ 638.187134] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 638.187134] env[62385]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 638.187134] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 638.187134] env[62385]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 638.187134] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 638.187134] env[62385]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 638.187134] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.187134] env[62385]: ERROR nova.compute.manager self.force_reraise() [ 638.187134] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.187134] env[62385]: ERROR nova.compute.manager raise self.value [ 638.187134] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 638.187134] env[62385]: ERROR nova.compute.manager updated_port = self._update_port( [ 638.187134] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.187134] env[62385]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 638.187562] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.187562] env[62385]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 638.187562] env[62385]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port be90e400-b8be-4151-9773-122efa78f030, please check neutron logs for more information. [ 638.187562] env[62385]: ERROR nova.compute.manager [ 638.187562] env[62385]: Traceback (most recent call last): [ 638.187562] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 638.187562] env[62385]: listener.cb(fileno) [ 638.187562] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 638.187562] env[62385]: result = function(*args, **kwargs) [ 638.187562] env[62385]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 638.187562] env[62385]: return func(*args, **kwargs) [ 638.187562] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 638.187562] env[62385]: raise e [ 638.187562] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 638.187562] env[62385]: nwinfo = self.network_api.allocate_for_instance( [ 638.187562] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 638.187562] env[62385]: created_port_ids = self._update_ports_for_instance( [ 638.187562] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 638.187562] env[62385]: with excutils.save_and_reraise_exception(): [ 638.187562] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.187562] env[62385]: self.force_reraise() [ 638.187562] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.187562] env[62385]: raise self.value [ 638.187562] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 638.187562] env[62385]: updated_port = self._update_port( [ 638.187562] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.187562] env[62385]: _ensure_no_port_binding_failure(port) [ 638.187562] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.187562] env[62385]: raise exception.PortBindingFailed(port_id=port['id']) [ 638.188467] env[62385]: nova.exception.PortBindingFailed: Binding failed for port be90e400-b8be-4151-9773-122efa78f030, please check neutron logs for more information. [ 638.188467] env[62385]: Removing descriptor: 19 [ 638.188467] env[62385]: ERROR nova.compute.manager [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] [instance: ad12e5b0-6854-4281-9155-e1660393829c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port be90e400-b8be-4151-9773-122efa78f030, please check neutron logs for more information. [ 638.188467] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] Traceback (most recent call last): [ 638.188467] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 638.188467] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] yield resources [ 638.188467] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 638.188467] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] self.driver.spawn(context, instance, image_meta, [ 638.188467] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 638.188467] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 638.188467] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 638.188467] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] vm_ref = self.build_virtual_machine(instance, [ 638.188888] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 638.188888] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] vif_infos = vmwarevif.get_vif_info(self._session, [ 638.188888] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 638.188888] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] for vif in network_info: [ 638.188888] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 638.188888] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] return self._sync_wrapper(fn, *args, **kwargs) [ 638.188888] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 638.188888] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] self.wait() [ 638.188888] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 638.188888] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] self[:] = self._gt.wait() [ 638.188888] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 638.188888] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] return self._exit_event.wait() [ 638.188888] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 638.189313] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] result = hub.switch() [ 638.189313] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 638.189313] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] return self.greenlet.switch() [ 638.189313] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 638.189313] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] result = function(*args, **kwargs) [ 638.189313] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 638.189313] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] return func(*args, **kwargs) [ 638.189313] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 638.189313] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] raise e [ 638.189313] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 638.189313] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] nwinfo = self.network_api.allocate_for_instance( [ 638.189313] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 638.189313] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] created_port_ids = self._update_ports_for_instance( [ 638.189698] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 638.189698] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] with excutils.save_and_reraise_exception(): [ 638.189698] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.189698] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] self.force_reraise() [ 638.189698] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.189698] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] raise self.value [ 638.189698] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 638.189698] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] updated_port = self._update_port( [ 638.189698] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.189698] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] _ensure_no_port_binding_failure(port) [ 638.189698] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.189698] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] raise exception.PortBindingFailed(port_id=port['id']) [ 638.190182] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] nova.exception.PortBindingFailed: Binding failed for port be90e400-b8be-4151-9773-122efa78f030, please check neutron logs for more information. [ 638.190182] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] [ 638.190182] env[62385]: INFO nova.compute.manager [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] [instance: ad12e5b0-6854-4281-9155-e1660393829c] Terminating instance [ 638.191280] env[62385]: DEBUG oslo_concurrency.lockutils [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] Acquiring lock "refresh_cache-ad12e5b0-6854-4281-9155-e1660393829c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.479171] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 608c13da-6625-417d-a7b3-10821ae638d8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 638.508337] env[62385]: DEBUG nova.network.neutron [req-4265e5d3-988d-41cb-b55b-5f2fa86e2674 req-1ea7599d-cc03-4013-a9b4-1dcca1816887 service nova] [instance: ad12e5b0-6854-4281-9155-e1660393829c] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 638.577917] env[62385]: DEBUG nova.network.neutron [req-4265e5d3-988d-41cb-b55b-5f2fa86e2674 req-1ea7599d-cc03-4013-a9b4-1dcca1816887 service nova] [instance: ad12e5b0-6854-4281-9155-e1660393829c] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.984025] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 6e0dd86b-d557-4e7a-8520-9ffbd291128b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 639.080971] env[62385]: DEBUG oslo_concurrency.lockutils [req-4265e5d3-988d-41cb-b55b-5f2fa86e2674 req-1ea7599d-cc03-4013-a9b4-1dcca1816887 service nova] Releasing lock "refresh_cache-ad12e5b0-6854-4281-9155-e1660393829c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 639.084151] env[62385]: DEBUG oslo_concurrency.lockutils [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] Acquired lock "refresh_cache-ad12e5b0-6854-4281-9155-e1660393829c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.084151] env[62385]: DEBUG nova.network.neutron [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] [instance: ad12e5b0-6854-4281-9155-e1660393829c] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 639.488015] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance ea0b59fe-8dd9-4893-87ff-54706eb64479 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 639.603941] env[62385]: DEBUG nova.network.neutron [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] [instance: ad12e5b0-6854-4281-9155-e1660393829c] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 639.654890] env[62385]: DEBUG nova.network.neutron [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] [instance: ad12e5b0-6854-4281-9155-e1660393829c] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.991701] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 639.994726] env[62385]: DEBUG nova.compute.manager [req-c4d877b3-d21f-4991-b936-2326aef47747 req-97bec2c3-597a-43e5-88e9-af573eba780e service nova] [instance: ad12e5b0-6854-4281-9155-e1660393829c] Received event network-vif-deleted-be90e400-b8be-4151-9773-122efa78f030 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 640.161266] env[62385]: DEBUG oslo_concurrency.lockutils [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] Releasing lock "refresh_cache-ad12e5b0-6854-4281-9155-e1660393829c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 640.161678] env[62385]: DEBUG nova.compute.manager [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] [instance: ad12e5b0-6854-4281-9155-e1660393829c] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 640.161910] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] [instance: ad12e5b0-6854-4281-9155-e1660393829c] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 640.162299] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0cdc5ef9-bdd2-4f46-9dfc-da5c149aaf00 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.170980] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a0257c4-55e4-4eaa-8b47-e0e8dd0881fa {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.193888] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] [instance: ad12e5b0-6854-4281-9155-e1660393829c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ad12e5b0-6854-4281-9155-e1660393829c could not be found. [ 640.194124] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] [instance: ad12e5b0-6854-4281-9155-e1660393829c] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 640.194305] env[62385]: INFO nova.compute.manager [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] [instance: ad12e5b0-6854-4281-9155-e1660393829c] Took 0.03 seconds to destroy the instance on the hypervisor. [ 640.194549] env[62385]: DEBUG oslo.service.loopingcall [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 640.194753] env[62385]: DEBUG nova.compute.manager [-] [instance: ad12e5b0-6854-4281-9155-e1660393829c] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 640.194849] env[62385]: DEBUG nova.network.neutron [-] [instance: ad12e5b0-6854-4281-9155-e1660393829c] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 640.218539] env[62385]: DEBUG nova.network.neutron [-] [instance: ad12e5b0-6854-4281-9155-e1660393829c] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 640.496527] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 1b459bac-e27d-4af5-bfc2-49898f5ced25 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 640.722379] env[62385]: DEBUG nova.network.neutron [-] [instance: ad12e5b0-6854-4281-9155-e1660393829c] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.999653] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance d3224596-4e38-4c9c-abd9-59a34bbf59c7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 641.225232] env[62385]: INFO nova.compute.manager [-] [instance: ad12e5b0-6854-4281-9155-e1660393829c] Took 1.03 seconds to deallocate network for instance. [ 641.227747] env[62385]: DEBUG nova.compute.claims [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] [instance: ad12e5b0-6854-4281-9155-e1660393829c] Aborting claim: {{(pid=62385) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 641.227933] env[62385]: DEBUG oslo_concurrency.lockutils [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.502214] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 642.005397] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 866451cc-96c5-433a-a903-9faa1eed538a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 642.509726] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 1ddb4d6d-3872-49db-bb40-e21721241e89 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 642.775819] env[62385]: DEBUG oslo_concurrency.lockutils [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Acquiring lock "65bb1d14-84b2-4e75-acdc-dc674a035101" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 642.776283] env[62385]: DEBUG oslo_concurrency.lockutils [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Lock "65bb1d14-84b2-4e75-acdc-dc674a035101" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 643.014214] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 643.518516] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 3e184125-28af-469b-83cc-4ab8859e0e2f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 644.021057] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 33387505-c576-488b-8c9c-b064fe81a7d7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 644.524230] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance f1d33401-35f5-4d79-abb4-26dc6faa784e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 645.030015] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance f50ddafe-f5a9-422b-b0f4-46c1b111dfbe has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 645.533496] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 646.037062] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance caca5304-0ad7-427c-812e-de925de63f2c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 646.540371] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance e97ff020-61f3-4947-bb82-5c039ad17747 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 647.045503] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 749e0976-2e2b-4764-865c-2e630f2edbd1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 647.548885] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 92057af7-28a3-4643-9cda-d3d868d0cfc9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 648.053869] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 42ecf594-dff7-4af3-ac56-5bbc45b7a192 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 648.557500] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance a5c4afc6-38a6-4815-8ec4-cc01c24489bf has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 649.060521] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance e73ecefc-fed3-4ccd-88d3-399a4b72bb4b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 649.060805] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=62385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 649.060953] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1856MB phys_disk=200GB used_disk=7GB total_vcpus=48 used_vcpus=7 pci_stats=[] {{(pid=62385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 649.382523] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b43de89-6118-4451-b2c8-c35799fc2888 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.391877] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6885b7fc-3199-4db3-8be4-6eb9581c2e08 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.421347] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12e0de7e-c999-4166-89f8-b05b57afb4bc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.428381] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d431df9a-5ec6-4f7e-bdba-df1c4f40c5c8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.441603] env[62385]: DEBUG nova.compute.provider_tree [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 649.944259] env[62385]: DEBUG nova.scheduler.client.report [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 650.449647] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62385) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 650.449896] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 14.547s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 650.450190] env[62385]: DEBUG oslo_concurrency.lockutils [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.654s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 651.339113] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e07f60c-ee95-42b6-bfdf-3371bf07181e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.346494] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a021e8fc-51bd-4be4-98e3-bb5a9f61c3f2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.376027] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1decf670-7ef8-41fe-b136-598626f2bb49 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.383387] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f53a4e25-2ffc-4beb-89e6-936f17027f95 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.396119] env[62385]: DEBUG nova.compute.provider_tree [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 651.899030] env[62385]: DEBUG nova.scheduler.client.report [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 652.403938] env[62385]: DEBUG oslo_concurrency.lockutils [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.954s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 652.404608] env[62385]: ERROR nova.compute.manager [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6eba3a70-eb34-4ef7-8145-12a451cc108b, please check neutron logs for more information. [ 652.404608] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Traceback (most recent call last): [ 652.404608] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 652.404608] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] self.driver.spawn(context, instance, image_meta, [ 652.404608] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 652.404608] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] self._vmops.spawn(context, instance, image_meta, injected_files, [ 652.404608] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 652.404608] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] vm_ref = self.build_virtual_machine(instance, [ 652.404608] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 652.404608] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] vif_infos = vmwarevif.get_vif_info(self._session, [ 652.404608] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 652.404955] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] for vif in network_info: [ 652.404955] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 652.404955] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] return self._sync_wrapper(fn, *args, **kwargs) [ 652.404955] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 652.404955] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] self.wait() [ 652.404955] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 652.404955] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] self[:] = self._gt.wait() [ 652.404955] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 652.404955] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] return self._exit_event.wait() [ 652.404955] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 652.404955] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] result = hub.switch() [ 652.404955] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 652.404955] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] return self.greenlet.switch() [ 652.405596] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 652.405596] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] result = function(*args, **kwargs) [ 652.405596] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 652.405596] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] return func(*args, **kwargs) [ 652.405596] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 652.405596] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] raise e [ 652.405596] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 652.405596] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] nwinfo = self.network_api.allocate_for_instance( [ 652.405596] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 652.405596] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] created_port_ids = self._update_ports_for_instance( [ 652.405596] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 652.405596] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] with excutils.save_and_reraise_exception(): [ 652.405596] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 652.406156] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] self.force_reraise() [ 652.406156] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 652.406156] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] raise self.value [ 652.406156] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 652.406156] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] updated_port = self._update_port( [ 652.406156] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 652.406156] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] _ensure_no_port_binding_failure(port) [ 652.406156] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 652.406156] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] raise exception.PortBindingFailed(port_id=port['id']) [ 652.406156] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] nova.exception.PortBindingFailed: Binding failed for port 6eba3a70-eb34-4ef7-8145-12a451cc108b, please check neutron logs for more information. [ 652.406156] env[62385]: ERROR nova.compute.manager [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] [ 652.406653] env[62385]: DEBUG nova.compute.utils [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Binding failed for port 6eba3a70-eb34-4ef7-8145-12a451cc108b, please check neutron logs for more information. {{(pid=62385) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 652.406708] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.199s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.408176] env[62385]: INFO nova.compute.claims [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: da94a435-b36c-42b2-8c84-564265872048] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 652.411026] env[62385]: DEBUG nova.compute.manager [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Build of instance 01d010dc-6b5b-470b-a387-f95a17600e19 was re-scheduled: Binding failed for port 6eba3a70-eb34-4ef7-8145-12a451cc108b, please check neutron logs for more information. {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 652.411328] env[62385]: DEBUG nova.compute.manager [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Unplugging VIFs for instance {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 652.411560] env[62385]: DEBUG oslo_concurrency.lockutils [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "refresh_cache-01d010dc-6b5b-470b-a387-f95a17600e19" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 652.411703] env[62385]: DEBUG oslo_concurrency.lockutils [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquired lock "refresh_cache-01d010dc-6b5b-470b-a387-f95a17600e19" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.411855] env[62385]: DEBUG nova.network.neutron [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 652.939467] env[62385]: DEBUG nova.network.neutron [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 653.129842] env[62385]: DEBUG nova.network.neutron [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.632984] env[62385]: DEBUG oslo_concurrency.lockutils [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Releasing lock "refresh_cache-01d010dc-6b5b-470b-a387-f95a17600e19" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 653.633309] env[62385]: DEBUG nova.compute.manager [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 653.633502] env[62385]: DEBUG nova.compute.manager [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 653.633671] env[62385]: DEBUG nova.network.neutron [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 653.652894] env[62385]: DEBUG nova.network.neutron [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 653.783412] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd76a1d4-869d-4cc5-84de-132e43be5baa {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.791027] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bde9d41a-9493-4689-b712-602b32ad173a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.819581] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1e05138-c5b2-4d37-a9bd-f3293a89e8e4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.826657] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d29df21-9435-4c30-b375-6d6a6a4341c9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.840560] env[62385]: DEBUG nova.compute.provider_tree [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 654.155728] env[62385]: DEBUG nova.network.neutron [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.343546] env[62385]: DEBUG nova.scheduler.client.report [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 654.658894] env[62385]: INFO nova.compute.manager [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: 01d010dc-6b5b-470b-a387-f95a17600e19] Took 1.02 seconds to deallocate network for instance. [ 654.848545] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.442s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 654.850034] env[62385]: DEBUG nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: da94a435-b36c-42b2-8c84-564265872048] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 654.852892] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.595s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 654.853337] env[62385]: INFO nova.compute.claims [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 655.358111] env[62385]: DEBUG nova.compute.utils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 655.365190] env[62385]: DEBUG nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: da94a435-b36c-42b2-8c84-564265872048] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 655.365190] env[62385]: DEBUG nova.network.neutron [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: da94a435-b36c-42b2-8c84-564265872048] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 655.425897] env[62385]: DEBUG nova.policy [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0b5753cd031043d1b22a799e3e72daad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f0965daa8a4e4ce8ab3ba0242ab77f74', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 655.686303] env[62385]: INFO nova.scheduler.client.report [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Deleted allocations for instance 01d010dc-6b5b-470b-a387-f95a17600e19 [ 655.843145] env[62385]: DEBUG nova.network.neutron [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: da94a435-b36c-42b2-8c84-564265872048] Successfully created port: 55745fbd-4c5f-410c-b72c-27ce94a14e33 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 655.866415] env[62385]: DEBUG nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: da94a435-b36c-42b2-8c84-564265872048] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 656.198292] env[62385]: DEBUG oslo_concurrency.lockutils [None req-014de420-383b-43db-b4d8-a787f0b2586f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "01d010dc-6b5b-470b-a387-f95a17600e19" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 102.367s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 656.256868] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-391bd23a-805a-4ba1-972a-46033403c9e9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.264462] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c50c9b16-ebca-4e86-a570-bd6f4712a437 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.295236] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36201826-ec57-45b6-9950-c8e40947b6a7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.302680] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d551249-c486-4593-acd1-a70d2402c9d0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.315521] env[62385]: DEBUG nova.compute.provider_tree [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 656.709242] env[62385]: DEBUG nova.compute.manager [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 656.787296] env[62385]: DEBUG nova.compute.manager [req-3c8890ec-5f61-4cf2-bb8f-eb48f604b19f req-52ecb866-d0e6-441a-a564-ca47b3eebe4a service nova] [instance: da94a435-b36c-42b2-8c84-564265872048] Received event network-changed-55745fbd-4c5f-410c-b72c-27ce94a14e33 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 656.787483] env[62385]: DEBUG nova.compute.manager [req-3c8890ec-5f61-4cf2-bb8f-eb48f604b19f req-52ecb866-d0e6-441a-a564-ca47b3eebe4a service nova] [instance: da94a435-b36c-42b2-8c84-564265872048] Refreshing instance network info cache due to event network-changed-55745fbd-4c5f-410c-b72c-27ce94a14e33. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 656.787693] env[62385]: DEBUG oslo_concurrency.lockutils [req-3c8890ec-5f61-4cf2-bb8f-eb48f604b19f req-52ecb866-d0e6-441a-a564-ca47b3eebe4a service nova] Acquiring lock "refresh_cache-da94a435-b36c-42b2-8c84-564265872048" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 656.787835] env[62385]: DEBUG oslo_concurrency.lockutils [req-3c8890ec-5f61-4cf2-bb8f-eb48f604b19f req-52ecb866-d0e6-441a-a564-ca47b3eebe4a service nova] Acquired lock "refresh_cache-da94a435-b36c-42b2-8c84-564265872048" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.788080] env[62385]: DEBUG nova.network.neutron [req-3c8890ec-5f61-4cf2-bb8f-eb48f604b19f req-52ecb866-d0e6-441a-a564-ca47b3eebe4a service nova] [instance: da94a435-b36c-42b2-8c84-564265872048] Refreshing network info cache for port 55745fbd-4c5f-410c-b72c-27ce94a14e33 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 656.820754] env[62385]: DEBUG nova.scheduler.client.report [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 656.875255] env[62385]: DEBUG nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: da94a435-b36c-42b2-8c84-564265872048] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 656.908633] env[62385]: DEBUG nova.virt.hardware [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 656.908867] env[62385]: DEBUG nova.virt.hardware [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 656.909036] env[62385]: DEBUG nova.virt.hardware [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 656.909260] env[62385]: DEBUG nova.virt.hardware [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 656.909416] env[62385]: DEBUG nova.virt.hardware [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 656.909567] env[62385]: DEBUG nova.virt.hardware [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 656.909773] env[62385]: DEBUG nova.virt.hardware [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 656.909932] env[62385]: DEBUG nova.virt.hardware [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 656.910115] env[62385]: DEBUG nova.virt.hardware [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 656.910279] env[62385]: DEBUG nova.virt.hardware [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 656.910454] env[62385]: DEBUG nova.virt.hardware [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 656.911344] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cddbe52-2109-417b-bcb1-2dd11051f008 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.920123] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ccbafa2-01c0-4f82-9271-b0feca614e89 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.977644] env[62385]: ERROR nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 55745fbd-4c5f-410c-b72c-27ce94a14e33, please check neutron logs for more information. [ 656.977644] env[62385]: ERROR nova.compute.manager Traceback (most recent call last): [ 656.977644] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.977644] env[62385]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 656.977644] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 656.977644] env[62385]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 656.977644] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 656.977644] env[62385]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 656.977644] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.977644] env[62385]: ERROR nova.compute.manager self.force_reraise() [ 656.977644] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.977644] env[62385]: ERROR nova.compute.manager raise self.value [ 656.977644] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 656.977644] env[62385]: ERROR nova.compute.manager updated_port = self._update_port( [ 656.977644] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.977644] env[62385]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 656.978122] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.978122] env[62385]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 656.978122] env[62385]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 55745fbd-4c5f-410c-b72c-27ce94a14e33, please check neutron logs for more information. [ 656.978122] env[62385]: ERROR nova.compute.manager [ 656.978122] env[62385]: Traceback (most recent call last): [ 656.978122] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 656.978122] env[62385]: listener.cb(fileno) [ 656.978122] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 656.978122] env[62385]: result = function(*args, **kwargs) [ 656.978122] env[62385]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 656.978122] env[62385]: return func(*args, **kwargs) [ 656.978122] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 656.978122] env[62385]: raise e [ 656.978122] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.978122] env[62385]: nwinfo = self.network_api.allocate_for_instance( [ 656.978122] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 656.978122] env[62385]: created_port_ids = self._update_ports_for_instance( [ 656.978122] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 656.978122] env[62385]: with excutils.save_and_reraise_exception(): [ 656.978122] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.978122] env[62385]: self.force_reraise() [ 656.978122] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.978122] env[62385]: raise self.value [ 656.978122] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 656.978122] env[62385]: updated_port = self._update_port( [ 656.978122] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.978122] env[62385]: _ensure_no_port_binding_failure(port) [ 656.978122] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.978122] env[62385]: raise exception.PortBindingFailed(port_id=port['id']) [ 656.979162] env[62385]: nova.exception.PortBindingFailed: Binding failed for port 55745fbd-4c5f-410c-b72c-27ce94a14e33, please check neutron logs for more information. [ 656.979162] env[62385]: Removing descriptor: 19 [ 656.979162] env[62385]: ERROR nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: da94a435-b36c-42b2-8c84-564265872048] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 55745fbd-4c5f-410c-b72c-27ce94a14e33, please check neutron logs for more information. [ 656.979162] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] Traceback (most recent call last): [ 656.979162] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 656.979162] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] yield resources [ 656.979162] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 656.979162] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] self.driver.spawn(context, instance, image_meta, [ 656.979162] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 656.979162] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] self._vmops.spawn(context, instance, image_meta, injected_files, [ 656.979162] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 656.979162] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] vm_ref = self.build_virtual_machine(instance, [ 656.979515] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 656.979515] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] vif_infos = vmwarevif.get_vif_info(self._session, [ 656.979515] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 656.979515] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] for vif in network_info: [ 656.979515] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 656.979515] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] return self._sync_wrapper(fn, *args, **kwargs) [ 656.979515] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 656.979515] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] self.wait() [ 656.979515] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 656.979515] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] self[:] = self._gt.wait() [ 656.979515] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 656.979515] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] return self._exit_event.wait() [ 656.979515] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 656.979904] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] result = hub.switch() [ 656.979904] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 656.979904] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] return self.greenlet.switch() [ 656.979904] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 656.979904] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] result = function(*args, **kwargs) [ 656.979904] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 656.979904] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] return func(*args, **kwargs) [ 656.979904] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 656.979904] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] raise e [ 656.979904] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.979904] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] nwinfo = self.network_api.allocate_for_instance( [ 656.979904] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 656.979904] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] created_port_ids = self._update_ports_for_instance( [ 656.980316] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 656.980316] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] with excutils.save_and_reraise_exception(): [ 656.980316] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.980316] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] self.force_reraise() [ 656.980316] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.980316] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] raise self.value [ 656.980316] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 656.980316] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] updated_port = self._update_port( [ 656.980316] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.980316] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] _ensure_no_port_binding_failure(port) [ 656.980316] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.980316] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] raise exception.PortBindingFailed(port_id=port['id']) [ 656.980664] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] nova.exception.PortBindingFailed: Binding failed for port 55745fbd-4c5f-410c-b72c-27ce94a14e33, please check neutron logs for more information. [ 656.980664] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] [ 656.980664] env[62385]: INFO nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: da94a435-b36c-42b2-8c84-564265872048] Terminating instance [ 656.981134] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Acquiring lock "refresh_cache-da94a435-b36c-42b2-8c84-564265872048" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 657.232007] env[62385]: DEBUG oslo_concurrency.lockutils [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.305959] env[62385]: DEBUG nova.network.neutron [req-3c8890ec-5f61-4cf2-bb8f-eb48f604b19f req-52ecb866-d0e6-441a-a564-ca47b3eebe4a service nova] [instance: da94a435-b36c-42b2-8c84-564265872048] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 657.322911] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.471s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 657.323454] env[62385]: DEBUG nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 657.326031] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.108s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.327455] env[62385]: INFO nova.compute.claims [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 657.408510] env[62385]: DEBUG nova.network.neutron [req-3c8890ec-5f61-4cf2-bb8f-eb48f604b19f req-52ecb866-d0e6-441a-a564-ca47b3eebe4a service nova] [instance: da94a435-b36c-42b2-8c84-564265872048] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.833429] env[62385]: DEBUG nova.compute.utils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 657.836583] env[62385]: DEBUG nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 657.836756] env[62385]: DEBUG nova.network.neutron [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 657.891719] env[62385]: DEBUG nova.policy [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0b5753cd031043d1b22a799e3e72daad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f0965daa8a4e4ce8ab3ba0242ab77f74', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 657.911564] env[62385]: DEBUG oslo_concurrency.lockutils [req-3c8890ec-5f61-4cf2-bb8f-eb48f604b19f req-52ecb866-d0e6-441a-a564-ca47b3eebe4a service nova] Releasing lock "refresh_cache-da94a435-b36c-42b2-8c84-564265872048" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 657.912311] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Acquired lock "refresh_cache-da94a435-b36c-42b2-8c84-564265872048" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.912311] env[62385]: DEBUG nova.network.neutron [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: da94a435-b36c-42b2-8c84-564265872048] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 658.224025] env[62385]: DEBUG nova.network.neutron [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Successfully created port: 4d8d9700-e4af-4437-88e3-fa5726600f49 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 658.337631] env[62385]: DEBUG nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 658.402189] env[62385]: DEBUG oslo_concurrency.lockutils [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "f8a98710-9d02-4aa3-b6a5-34b7060ad62d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 658.403031] env[62385]: DEBUG oslo_concurrency.lockutils [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "f8a98710-9d02-4aa3-b6a5-34b7060ad62d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 658.433153] env[62385]: DEBUG nova.network.neutron [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: da94a435-b36c-42b2-8c84-564265872048] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 658.507716] env[62385]: DEBUG nova.network.neutron [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: da94a435-b36c-42b2-8c84-564265872048] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.728020] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79252483-032f-4c7d-b374-572871189038 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.734653] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c61abfb-14d5-494a-966a-63d9843ae2f6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.765438] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e36b90a-9247-4e41-85dd-9ef75a9425d7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.773162] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4ee64bb-3a91-4dae-8778-18199fb0e21a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.790420] env[62385]: DEBUG nova.compute.provider_tree [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 658.816821] env[62385]: DEBUG nova.compute.manager [req-d1ec616c-b78d-4849-a04a-64fe938461a0 req-2b580c49-ae9f-49fb-a8ac-949251dcf889 service nova] [instance: da94a435-b36c-42b2-8c84-564265872048] Received event network-vif-deleted-55745fbd-4c5f-410c-b72c-27ce94a14e33 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 659.010557] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Releasing lock "refresh_cache-da94a435-b36c-42b2-8c84-564265872048" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 659.011031] env[62385]: DEBUG nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: da94a435-b36c-42b2-8c84-564265872048] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 659.011251] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: da94a435-b36c-42b2-8c84-564265872048] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 659.011550] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8964acaa-a29f-4870-8e6e-fbd43adb3d45 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.020818] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e42c5f7-13c4-4da0-b410-08d86240f89d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.043439] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: da94a435-b36c-42b2-8c84-564265872048] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance da94a435-b36c-42b2-8c84-564265872048 could not be found. [ 659.043655] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: da94a435-b36c-42b2-8c84-564265872048] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 659.043833] env[62385]: INFO nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: da94a435-b36c-42b2-8c84-564265872048] Took 0.03 seconds to destroy the instance on the hypervisor. [ 659.044082] env[62385]: DEBUG oslo.service.loopingcall [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 659.044300] env[62385]: DEBUG nova.compute.manager [-] [instance: da94a435-b36c-42b2-8c84-564265872048] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 659.044391] env[62385]: DEBUG nova.network.neutron [-] [instance: da94a435-b36c-42b2-8c84-564265872048] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 659.064101] env[62385]: DEBUG nova.network.neutron [-] [instance: da94a435-b36c-42b2-8c84-564265872048] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 659.295024] env[62385]: DEBUG nova.scheduler.client.report [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 659.308408] env[62385]: ERROR nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4d8d9700-e4af-4437-88e3-fa5726600f49, please check neutron logs for more information. [ 659.308408] env[62385]: ERROR nova.compute.manager Traceback (most recent call last): [ 659.308408] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 659.308408] env[62385]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 659.308408] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 659.308408] env[62385]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 659.308408] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 659.308408] env[62385]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 659.308408] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 659.308408] env[62385]: ERROR nova.compute.manager self.force_reraise() [ 659.308408] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 659.308408] env[62385]: ERROR nova.compute.manager raise self.value [ 659.308408] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 659.308408] env[62385]: ERROR nova.compute.manager updated_port = self._update_port( [ 659.308408] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 659.308408] env[62385]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 659.308932] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 659.308932] env[62385]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 659.308932] env[62385]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4d8d9700-e4af-4437-88e3-fa5726600f49, please check neutron logs for more information. [ 659.308932] env[62385]: ERROR nova.compute.manager [ 659.308932] env[62385]: Traceback (most recent call last): [ 659.308932] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 659.308932] env[62385]: listener.cb(fileno) [ 659.308932] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 659.308932] env[62385]: result = function(*args, **kwargs) [ 659.308932] env[62385]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 659.308932] env[62385]: return func(*args, **kwargs) [ 659.308932] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 659.308932] env[62385]: raise e [ 659.308932] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 659.308932] env[62385]: nwinfo = self.network_api.allocate_for_instance( [ 659.308932] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 659.308932] env[62385]: created_port_ids = self._update_ports_for_instance( [ 659.308932] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 659.308932] env[62385]: with excutils.save_and_reraise_exception(): [ 659.308932] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 659.308932] env[62385]: self.force_reraise() [ 659.308932] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 659.308932] env[62385]: raise self.value [ 659.308932] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 659.308932] env[62385]: updated_port = self._update_port( [ 659.308932] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 659.308932] env[62385]: _ensure_no_port_binding_failure(port) [ 659.308932] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 659.308932] env[62385]: raise exception.PortBindingFailed(port_id=port['id']) [ 659.309755] env[62385]: nova.exception.PortBindingFailed: Binding failed for port 4d8d9700-e4af-4437-88e3-fa5726600f49, please check neutron logs for more information. [ 659.309755] env[62385]: Removing descriptor: 19 [ 659.349595] env[62385]: DEBUG nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 659.383975] env[62385]: DEBUG nova.virt.hardware [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 659.384274] env[62385]: DEBUG nova.virt.hardware [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 659.384499] env[62385]: DEBUG nova.virt.hardware [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 659.384836] env[62385]: DEBUG nova.virt.hardware [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 659.385115] env[62385]: DEBUG nova.virt.hardware [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 659.385340] env[62385]: DEBUG nova.virt.hardware [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 659.385655] env[62385]: DEBUG nova.virt.hardware [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 659.385888] env[62385]: DEBUG nova.virt.hardware [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 659.386153] env[62385]: DEBUG nova.virt.hardware [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 659.386398] env[62385]: DEBUG nova.virt.hardware [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 659.386640] env[62385]: DEBUG nova.virt.hardware [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 659.388061] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ea1486c-5101-470b-b557-a60a927f0f6a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.399173] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b1afc06-492f-4e36-8786-5c86c1fbcda3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.415585] env[62385]: ERROR nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4d8d9700-e4af-4437-88e3-fa5726600f49, please check neutron logs for more information. [ 659.415585] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Traceback (most recent call last): [ 659.415585] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 659.415585] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] yield resources [ 659.415585] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 659.415585] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] self.driver.spawn(context, instance, image_meta, [ 659.415585] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 659.415585] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] self._vmops.spawn(context, instance, image_meta, injected_files, [ 659.415585] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 659.415585] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] vm_ref = self.build_virtual_machine(instance, [ 659.415585] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 659.415910] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] vif_infos = vmwarevif.get_vif_info(self._session, [ 659.415910] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 659.415910] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] for vif in network_info: [ 659.415910] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 659.415910] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] return self._sync_wrapper(fn, *args, **kwargs) [ 659.415910] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 659.415910] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] self.wait() [ 659.415910] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 659.415910] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] self[:] = self._gt.wait() [ 659.415910] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 659.415910] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] return self._exit_event.wait() [ 659.415910] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 659.415910] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] current.throw(*self._exc) [ 659.416249] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 659.416249] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] result = function(*args, **kwargs) [ 659.416249] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 659.416249] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] return func(*args, **kwargs) [ 659.416249] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 659.416249] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] raise e [ 659.416249] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 659.416249] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] nwinfo = self.network_api.allocate_for_instance( [ 659.416249] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 659.416249] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] created_port_ids = self._update_ports_for_instance( [ 659.416249] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 659.416249] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] with excutils.save_and_reraise_exception(): [ 659.416249] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 659.416689] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] self.force_reraise() [ 659.416689] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 659.416689] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] raise self.value [ 659.416689] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 659.416689] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] updated_port = self._update_port( [ 659.416689] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 659.416689] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] _ensure_no_port_binding_failure(port) [ 659.416689] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 659.416689] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] raise exception.PortBindingFailed(port_id=port['id']) [ 659.416689] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] nova.exception.PortBindingFailed: Binding failed for port 4d8d9700-e4af-4437-88e3-fa5726600f49, please check neutron logs for more information. [ 659.416689] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] [ 659.416689] env[62385]: INFO nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Terminating instance [ 659.418950] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Acquiring lock "refresh_cache-7b159d98-9fd2-4eca-9755-cd2cac502798" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 659.419051] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Acquired lock "refresh_cache-7b159d98-9fd2-4eca-9755-cd2cac502798" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.419193] env[62385]: DEBUG nova.network.neutron [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 659.568513] env[62385]: DEBUG nova.network.neutron [-] [instance: da94a435-b36c-42b2-8c84-564265872048] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.797984] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.472s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 659.798885] env[62385]: DEBUG nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 659.801044] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 33.125s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 659.945573] env[62385]: DEBUG nova.network.neutron [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 660.040841] env[62385]: DEBUG nova.network.neutron [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.070560] env[62385]: INFO nova.compute.manager [-] [instance: da94a435-b36c-42b2-8c84-564265872048] Took 1.03 seconds to deallocate network for instance. [ 660.072862] env[62385]: DEBUG nova.compute.claims [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: da94a435-b36c-42b2-8c84-564265872048] Aborting claim: {{(pid=62385) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 660.073058] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 660.308931] env[62385]: DEBUG nova.compute.utils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 660.310715] env[62385]: DEBUG nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 660.311136] env[62385]: DEBUG nova.network.neutron [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 660.347870] env[62385]: DEBUG nova.policy [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0b5753cd031043d1b22a799e3e72daad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f0965daa8a4e4ce8ab3ba0242ab77f74', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 660.543460] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Releasing lock "refresh_cache-7b159d98-9fd2-4eca-9755-cd2cac502798" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 660.543869] env[62385]: DEBUG nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 660.544075] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 660.544378] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8a75324a-3938-4dec-a9b4-2f3bba906576 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.553420] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e35d7a82-2d76-4f81-b458-605d30f41e92 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.577881] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7b159d98-9fd2-4eca-9755-cd2cac502798 could not be found. [ 660.578129] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 660.578332] env[62385]: INFO nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Took 0.03 seconds to destroy the instance on the hypervisor. [ 660.578611] env[62385]: DEBUG oslo.service.loopingcall [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 660.580919] env[62385]: DEBUG nova.compute.manager [-] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 660.581063] env[62385]: DEBUG nova.network.neutron [-] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 660.599625] env[62385]: DEBUG nova.network.neutron [-] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 660.702765] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c56325f-c995-4a5b-82a4-53753148428e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.710749] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33509e7e-ccf2-43ca-84f7-6d9f0766ddc6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.742010] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b2944eb-0aa5-4a35-a721-c384404fabba {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.749659] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c13d41eb-054e-4606-a607-d7974ababcdf {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.764401] env[62385]: DEBUG nova.compute.provider_tree [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 660.813731] env[62385]: DEBUG nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 660.818978] env[62385]: DEBUG nova.network.neutron [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Successfully created port: 6f8b5530-e4f3-4741-a192-0e956321a364 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 660.988756] env[62385]: DEBUG nova.compute.manager [req-a9698158-dba4-4f76-99ea-c123faae8864 req-29d9c705-9a83-4ec7-99ae-59194cc825ed service nova] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Received event network-changed-4d8d9700-e4af-4437-88e3-fa5726600f49 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 660.988964] env[62385]: DEBUG nova.compute.manager [req-a9698158-dba4-4f76-99ea-c123faae8864 req-29d9c705-9a83-4ec7-99ae-59194cc825ed service nova] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Refreshing instance network info cache due to event network-changed-4d8d9700-e4af-4437-88e3-fa5726600f49. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 660.989203] env[62385]: DEBUG oslo_concurrency.lockutils [req-a9698158-dba4-4f76-99ea-c123faae8864 req-29d9c705-9a83-4ec7-99ae-59194cc825ed service nova] Acquiring lock "refresh_cache-7b159d98-9fd2-4eca-9755-cd2cac502798" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.989375] env[62385]: DEBUG oslo_concurrency.lockutils [req-a9698158-dba4-4f76-99ea-c123faae8864 req-29d9c705-9a83-4ec7-99ae-59194cc825ed service nova] Acquired lock "refresh_cache-7b159d98-9fd2-4eca-9755-cd2cac502798" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.989548] env[62385]: DEBUG nova.network.neutron [req-a9698158-dba4-4f76-99ea-c123faae8864 req-29d9c705-9a83-4ec7-99ae-59194cc825ed service nova] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Refreshing network info cache for port 4d8d9700-e4af-4437-88e3-fa5726600f49 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 661.102011] env[62385]: DEBUG nova.network.neutron [-] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.267214] env[62385]: DEBUG nova.scheduler.client.report [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 661.511561] env[62385]: DEBUG nova.network.neutron [req-a9698158-dba4-4f76-99ea-c123faae8864 req-29d9c705-9a83-4ec7-99ae-59194cc825ed service nova] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 661.591312] env[62385]: DEBUG nova.network.neutron [req-a9698158-dba4-4f76-99ea-c123faae8864 req-29d9c705-9a83-4ec7-99ae-59194cc825ed service nova] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.606064] env[62385]: INFO nova.compute.manager [-] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Took 1.02 seconds to deallocate network for instance. [ 661.608159] env[62385]: DEBUG nova.compute.claims [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Aborting claim: {{(pid=62385) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 661.608334] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 661.771899] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.971s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 661.772555] env[62385]: ERROR nova.compute.manager [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port baf9e6b1-3e80-4e7f-9ed6-e6cdabd7320a, please check neutron logs for more information. [ 661.772555] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Traceback (most recent call last): [ 661.772555] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 661.772555] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] self.driver.spawn(context, instance, image_meta, [ 661.772555] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 661.772555] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 661.772555] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 661.772555] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] vm_ref = self.build_virtual_machine(instance, [ 661.772555] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 661.772555] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] vif_infos = vmwarevif.get_vif_info(self._session, [ 661.772555] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 661.772972] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] for vif in network_info: [ 661.772972] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 661.772972] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] return self._sync_wrapper(fn, *args, **kwargs) [ 661.772972] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 661.772972] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] self.wait() [ 661.772972] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 661.772972] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] self[:] = self._gt.wait() [ 661.772972] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 661.772972] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] return self._exit_event.wait() [ 661.772972] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 661.772972] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] result = hub.switch() [ 661.772972] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 661.772972] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] return self.greenlet.switch() [ 661.773376] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 661.773376] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] result = function(*args, **kwargs) [ 661.773376] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 661.773376] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] return func(*args, **kwargs) [ 661.773376] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 661.773376] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] raise e [ 661.773376] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 661.773376] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] nwinfo = self.network_api.allocate_for_instance( [ 661.773376] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 661.773376] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] created_port_ids = self._update_ports_for_instance( [ 661.773376] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 661.773376] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] with excutils.save_and_reraise_exception(): [ 661.773376] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 661.774031] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] self.force_reraise() [ 661.774031] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 661.774031] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] raise self.value [ 661.774031] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 661.774031] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] updated_port = self._update_port( [ 661.774031] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 661.774031] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] _ensure_no_port_binding_failure(port) [ 661.774031] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 661.774031] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] raise exception.PortBindingFailed(port_id=port['id']) [ 661.774031] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] nova.exception.PortBindingFailed: Binding failed for port baf9e6b1-3e80-4e7f-9ed6-e6cdabd7320a, please check neutron logs for more information. [ 661.774031] env[62385]: ERROR nova.compute.manager [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] [ 661.774465] env[62385]: DEBUG nova.compute.utils [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Binding failed for port baf9e6b1-3e80-4e7f-9ed6-e6cdabd7320a, please check neutron logs for more information. {{(pid=62385) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 661.774520] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 32.546s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 661.777999] env[62385]: DEBUG nova.compute.manager [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Build of instance cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e was re-scheduled: Binding failed for port baf9e6b1-3e80-4e7f-9ed6-e6cdabd7320a, please check neutron logs for more information. {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 661.778472] env[62385]: DEBUG nova.compute.manager [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Unplugging VIFs for instance {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 661.778696] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] Acquiring lock "refresh_cache-cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 661.778842] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] Acquired lock "refresh_cache-cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.778999] env[62385]: DEBUG nova.network.neutron [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 661.824094] env[62385]: DEBUG nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 661.851251] env[62385]: DEBUG nova.virt.hardware [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 661.851491] env[62385]: DEBUG nova.virt.hardware [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 661.851739] env[62385]: DEBUG nova.virt.hardware [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 661.851939] env[62385]: DEBUG nova.virt.hardware [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 661.852119] env[62385]: DEBUG nova.virt.hardware [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 661.852278] env[62385]: DEBUG nova.virt.hardware [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 661.852485] env[62385]: DEBUG nova.virt.hardware [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 661.852641] env[62385]: DEBUG nova.virt.hardware [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 661.853094] env[62385]: DEBUG nova.virt.hardware [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 661.853094] env[62385]: DEBUG nova.virt.hardware [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 661.853203] env[62385]: DEBUG nova.virt.hardware [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 661.854238] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4acf050-9411-4c07-8da2-64aa8510b9a4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.862146] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6c0d65f-0bd8-41eb-9e27-b2bdd04b2b83 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.920735] env[62385]: ERROR nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6f8b5530-e4f3-4741-a192-0e956321a364, please check neutron logs for more information. [ 661.920735] env[62385]: ERROR nova.compute.manager Traceback (most recent call last): [ 661.920735] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 661.920735] env[62385]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 661.920735] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 661.920735] env[62385]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 661.920735] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 661.920735] env[62385]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 661.920735] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 661.920735] env[62385]: ERROR nova.compute.manager self.force_reraise() [ 661.920735] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 661.920735] env[62385]: ERROR nova.compute.manager raise self.value [ 661.920735] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 661.920735] env[62385]: ERROR nova.compute.manager updated_port = self._update_port( [ 661.920735] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 661.920735] env[62385]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 661.921221] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 661.921221] env[62385]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 661.921221] env[62385]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6f8b5530-e4f3-4741-a192-0e956321a364, please check neutron logs for more information. [ 661.921221] env[62385]: ERROR nova.compute.manager [ 661.921221] env[62385]: Traceback (most recent call last): [ 661.921221] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 661.921221] env[62385]: listener.cb(fileno) [ 661.921221] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 661.921221] env[62385]: result = function(*args, **kwargs) [ 661.921221] env[62385]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 661.921221] env[62385]: return func(*args, **kwargs) [ 661.921221] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 661.921221] env[62385]: raise e [ 661.921221] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 661.921221] env[62385]: nwinfo = self.network_api.allocate_for_instance( [ 661.921221] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 661.921221] env[62385]: created_port_ids = self._update_ports_for_instance( [ 661.921221] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 661.921221] env[62385]: with excutils.save_and_reraise_exception(): [ 661.921221] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 661.921221] env[62385]: self.force_reraise() [ 661.921221] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 661.921221] env[62385]: raise self.value [ 661.921221] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 661.921221] env[62385]: updated_port = self._update_port( [ 661.921221] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 661.921221] env[62385]: _ensure_no_port_binding_failure(port) [ 661.921221] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 661.921221] env[62385]: raise exception.PortBindingFailed(port_id=port['id']) [ 661.922136] env[62385]: nova.exception.PortBindingFailed: Binding failed for port 6f8b5530-e4f3-4741-a192-0e956321a364, please check neutron logs for more information. [ 661.922136] env[62385]: Removing descriptor: 19 [ 661.922136] env[62385]: ERROR nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6f8b5530-e4f3-4741-a192-0e956321a364, please check neutron logs for more information. [ 661.922136] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Traceback (most recent call last): [ 661.922136] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 661.922136] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] yield resources [ 661.922136] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 661.922136] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] self.driver.spawn(context, instance, image_meta, [ 661.922136] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 661.922136] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 661.922136] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 661.922136] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] vm_ref = self.build_virtual_machine(instance, [ 661.922509] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 661.922509] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] vif_infos = vmwarevif.get_vif_info(self._session, [ 661.922509] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 661.922509] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] for vif in network_info: [ 661.922509] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 661.922509] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] return self._sync_wrapper(fn, *args, **kwargs) [ 661.922509] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 661.922509] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] self.wait() [ 661.922509] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 661.922509] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] self[:] = self._gt.wait() [ 661.922509] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 661.922509] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] return self._exit_event.wait() [ 661.922509] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 661.922890] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] result = hub.switch() [ 661.922890] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 661.922890] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] return self.greenlet.switch() [ 661.922890] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 661.922890] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] result = function(*args, **kwargs) [ 661.922890] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 661.922890] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] return func(*args, **kwargs) [ 661.922890] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 661.922890] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] raise e [ 661.922890] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 661.922890] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] nwinfo = self.network_api.allocate_for_instance( [ 661.922890] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 661.922890] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] created_port_ids = self._update_ports_for_instance( [ 661.923275] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 661.923275] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] with excutils.save_and_reraise_exception(): [ 661.923275] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 661.923275] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] self.force_reraise() [ 661.923275] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 661.923275] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] raise self.value [ 661.923275] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 661.923275] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] updated_port = self._update_port( [ 661.923275] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 661.923275] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] _ensure_no_port_binding_failure(port) [ 661.923275] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 661.923275] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] raise exception.PortBindingFailed(port_id=port['id']) [ 661.923616] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] nova.exception.PortBindingFailed: Binding failed for port 6f8b5530-e4f3-4741-a192-0e956321a364, please check neutron logs for more information. [ 661.923616] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] [ 661.923616] env[62385]: INFO nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Terminating instance [ 661.923956] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Acquiring lock "refresh_cache-608c13da-6625-417d-a7b3-10821ae638d8" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 661.924133] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Acquired lock "refresh_cache-608c13da-6625-417d-a7b3-10821ae638d8" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.924298] env[62385]: DEBUG nova.network.neutron [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 662.093830] env[62385]: DEBUG oslo_concurrency.lockutils [req-a9698158-dba4-4f76-99ea-c123faae8864 req-29d9c705-9a83-4ec7-99ae-59194cc825ed service nova] Releasing lock "refresh_cache-7b159d98-9fd2-4eca-9755-cd2cac502798" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 662.094554] env[62385]: DEBUG nova.compute.manager [req-a9698158-dba4-4f76-99ea-c123faae8864 req-29d9c705-9a83-4ec7-99ae-59194cc825ed service nova] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Received event network-vif-deleted-4d8d9700-e4af-4437-88e3-fa5726600f49 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 662.298456] env[62385]: DEBUG nova.network.neutron [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 662.418663] env[62385]: DEBUG nova.network.neutron [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.444023] env[62385]: DEBUG nova.network.neutron [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 662.535386] env[62385]: DEBUG nova.network.neutron [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.643620] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a43607ca-9289-4767-92c3-3ce68b89c560 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.650878] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84d0035d-3824-42fb-98f6-dc10de7bfa43 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.679978] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-702af6d3-0202-49de-a184-704f65bcbe66 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.687302] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10efd271-6d24-469b-851a-80786f6d8f3d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.700125] env[62385]: DEBUG nova.compute.provider_tree [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 662.921824] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] Releasing lock "refresh_cache-cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 662.922051] env[62385]: DEBUG nova.compute.manager [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 662.922246] env[62385]: DEBUG nova.compute.manager [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 662.922415] env[62385]: DEBUG nova.network.neutron [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 662.941924] env[62385]: DEBUG nova.network.neutron [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 663.016856] env[62385]: DEBUG nova.compute.manager [req-88def5dc-f843-44a4-ab36-284ea4cbe138 req-36945999-e6ba-4487-9b27-8571853c2174 service nova] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Received event network-changed-6f8b5530-e4f3-4741-a192-0e956321a364 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 663.016856] env[62385]: DEBUG nova.compute.manager [req-88def5dc-f843-44a4-ab36-284ea4cbe138 req-36945999-e6ba-4487-9b27-8571853c2174 service nova] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Refreshing instance network info cache due to event network-changed-6f8b5530-e4f3-4741-a192-0e956321a364. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 663.016856] env[62385]: DEBUG oslo_concurrency.lockutils [req-88def5dc-f843-44a4-ab36-284ea4cbe138 req-36945999-e6ba-4487-9b27-8571853c2174 service nova] Acquiring lock "refresh_cache-608c13da-6625-417d-a7b3-10821ae638d8" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 663.038574] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Releasing lock "refresh_cache-608c13da-6625-417d-a7b3-10821ae638d8" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 663.038987] env[62385]: DEBUG nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 663.039210] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 663.039837] env[62385]: DEBUG oslo_concurrency.lockutils [req-88def5dc-f843-44a4-ab36-284ea4cbe138 req-36945999-e6ba-4487-9b27-8571853c2174 service nova] Acquired lock "refresh_cache-608c13da-6625-417d-a7b3-10821ae638d8" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.040029] env[62385]: DEBUG nova.network.neutron [req-88def5dc-f843-44a4-ab36-284ea4cbe138 req-36945999-e6ba-4487-9b27-8571853c2174 service nova] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Refreshing network info cache for port 6f8b5530-e4f3-4741-a192-0e956321a364 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 663.041074] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-85227a90-14b8-4d42-b207-4fc5b862f3cc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.050501] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e42da678-e044-4490-a377-072a08772996 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.073976] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 608c13da-6625-417d-a7b3-10821ae638d8 could not be found. [ 663.075026] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 663.075026] env[62385]: INFO nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Took 0.04 seconds to destroy the instance on the hypervisor. [ 663.075026] env[62385]: DEBUG oslo.service.loopingcall [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 663.075026] env[62385]: DEBUG nova.compute.manager [-] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 663.075026] env[62385]: DEBUG nova.network.neutron [-] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 663.093928] env[62385]: DEBUG nova.network.neutron [-] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 663.202950] env[62385]: DEBUG nova.scheduler.client.report [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 663.444236] env[62385]: DEBUG nova.network.neutron [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.558796] env[62385]: DEBUG nova.network.neutron [req-88def5dc-f843-44a4-ab36-284ea4cbe138 req-36945999-e6ba-4487-9b27-8571853c2174 service nova] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 663.596021] env[62385]: DEBUG nova.network.neutron [-] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.631195] env[62385]: DEBUG nova.network.neutron [req-88def5dc-f843-44a4-ab36-284ea4cbe138 req-36945999-e6ba-4487-9b27-8571853c2174 service nova] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.707475] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.933s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 663.708148] env[62385]: ERROR nova.compute.manager [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f51c0b92-691e-4172-8fa3-d6d28cc46bcf, please check neutron logs for more information. [ 663.708148] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Traceback (most recent call last): [ 663.708148] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 663.708148] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] self.driver.spawn(context, instance, image_meta, [ 663.708148] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 663.708148] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 663.708148] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 663.708148] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] vm_ref = self.build_virtual_machine(instance, [ 663.708148] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 663.708148] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] vif_infos = vmwarevif.get_vif_info(self._session, [ 663.708148] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 663.708535] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] for vif in network_info: [ 663.708535] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 663.708535] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] return self._sync_wrapper(fn, *args, **kwargs) [ 663.708535] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 663.708535] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] self.wait() [ 663.708535] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 663.708535] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] self[:] = self._gt.wait() [ 663.708535] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 663.708535] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] return self._exit_event.wait() [ 663.708535] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 663.708535] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] result = hub.switch() [ 663.708535] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 663.708535] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] return self.greenlet.switch() [ 663.708907] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 663.708907] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] result = function(*args, **kwargs) [ 663.708907] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 663.708907] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] return func(*args, **kwargs) [ 663.708907] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 663.708907] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] raise e [ 663.708907] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 663.708907] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] nwinfo = self.network_api.allocate_for_instance( [ 663.708907] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 663.708907] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] created_port_ids = self._update_ports_for_instance( [ 663.708907] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 663.708907] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] with excutils.save_and_reraise_exception(): [ 663.708907] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 663.709313] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] self.force_reraise() [ 663.709313] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 663.709313] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] raise self.value [ 663.709313] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 663.709313] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] updated_port = self._update_port( [ 663.709313] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 663.709313] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] _ensure_no_port_binding_failure(port) [ 663.709313] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 663.709313] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] raise exception.PortBindingFailed(port_id=port['id']) [ 663.709313] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] nova.exception.PortBindingFailed: Binding failed for port f51c0b92-691e-4172-8fa3-d6d28cc46bcf, please check neutron logs for more information. [ 663.709313] env[62385]: ERROR nova.compute.manager [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] [ 663.709636] env[62385]: DEBUG nova.compute.utils [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Binding failed for port f51c0b92-691e-4172-8fa3-d6d28cc46bcf, please check neutron logs for more information. {{(pid=62385) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 663.709938] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 31.584s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 663.710128] env[62385]: DEBUG nova.objects.instance [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62385) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 663.712534] env[62385]: DEBUG nova.compute.manager [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Build of instance 70abd86c-aa63-4ccb-b185-65bf4977dfdf was re-scheduled: Binding failed for port f51c0b92-691e-4172-8fa3-d6d28cc46bcf, please check neutron logs for more information. {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 663.713157] env[62385]: DEBUG nova.compute.manager [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Unplugging VIFs for instance {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 663.713237] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Acquiring lock "refresh_cache-70abd86c-aa63-4ccb-b185-65bf4977dfdf" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 663.713313] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Acquired lock "refresh_cache-70abd86c-aa63-4ccb-b185-65bf4977dfdf" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.713459] env[62385]: DEBUG nova.network.neutron [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 663.953028] env[62385]: INFO nova.compute.manager [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] [instance: cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e] Took 1.03 seconds to deallocate network for instance. [ 664.098497] env[62385]: INFO nova.compute.manager [-] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Took 1.02 seconds to deallocate network for instance. [ 664.102433] env[62385]: DEBUG nova.compute.claims [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Aborting claim: {{(pid=62385) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 664.102433] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.133547] env[62385]: DEBUG oslo_concurrency.lockutils [req-88def5dc-f843-44a4-ab36-284ea4cbe138 req-36945999-e6ba-4487-9b27-8571853c2174 service nova] Releasing lock "refresh_cache-608c13da-6625-417d-a7b3-10821ae638d8" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 664.133817] env[62385]: DEBUG nova.compute.manager [req-88def5dc-f843-44a4-ab36-284ea4cbe138 req-36945999-e6ba-4487-9b27-8571853c2174 service nova] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Received event network-vif-deleted-6f8b5530-e4f3-4741-a192-0e956321a364 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 664.232818] env[62385]: DEBUG nova.network.neutron [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 664.304077] env[62385]: DEBUG nova.network.neutron [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.723434] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4e0cb1d6-5b2a-4231-84ed-434495b79de0 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 664.724587] env[62385]: DEBUG oslo_concurrency.lockutils [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.201s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.726118] env[62385]: INFO nova.compute.claims [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 664.806459] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Releasing lock "refresh_cache-70abd86c-aa63-4ccb-b185-65bf4977dfdf" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 664.806690] env[62385]: DEBUG nova.compute.manager [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 664.806853] env[62385]: DEBUG nova.compute.manager [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 664.807055] env[62385]: DEBUG nova.network.neutron [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 664.822021] env[62385]: DEBUG nova.network.neutron [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 664.980413] env[62385]: INFO nova.scheduler.client.report [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] Deleted allocations for instance cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e [ 665.324486] env[62385]: DEBUG nova.network.neutron [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.492590] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6037e4fb-3036-4d15-8da2-e6d1746981c2 tempest-ServersTestManualDisk-899014057 tempest-ServersTestManualDisk-899014057-project-member] Lock "cc2bce8f-0c0c-4cbc-a31c-c8ec2992a76e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 111.649s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 665.826917] env[62385]: INFO nova.compute.manager [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] [instance: 70abd86c-aa63-4ccb-b185-65bf4977dfdf] Took 1.02 seconds to deallocate network for instance. [ 665.997069] env[62385]: DEBUG nova.compute.manager [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 666.134874] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e508236-3e5c-4b58-9f96-60558db781df {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.143729] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5d7b211-3379-4a17-95e3-95ebd72642ee {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.176508] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75dc0ece-6c9f-4c91-a67a-6019167c61ce {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.184902] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58521535-8eb9-406c-94a5-df9933c2b3fd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.199202] env[62385]: DEBUG nova.compute.provider_tree [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 666.513983] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 666.701694] env[62385]: DEBUG nova.scheduler.client.report [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 666.857019] env[62385]: INFO nova.scheduler.client.report [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Deleted allocations for instance 70abd86c-aa63-4ccb-b185-65bf4977dfdf [ 667.209017] env[62385]: DEBUG oslo_concurrency.lockutils [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.482s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 667.209017] env[62385]: DEBUG nova.compute.manager [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 667.213076] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.174s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 667.363054] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5b3f261c-91e1-46d5-a95a-72b89b4e6c9f tempest-DeleteServersAdminTestJSON-664582793 tempest-DeleteServersAdminTestJSON-664582793-project-member] Lock "70abd86c-aa63-4ccb-b185-65bf4977dfdf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 112.903s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 667.715643] env[62385]: DEBUG nova.compute.utils [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 667.723968] env[62385]: DEBUG nova.compute.manager [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 667.723968] env[62385]: DEBUG nova.network.neutron [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 667.818506] env[62385]: DEBUG nova.policy [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'effe2cfd64d44e32a258588162812cba', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '83d39da6eb0541fea18f25acc9df3eef', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 667.868732] env[62385]: DEBUG nova.compute.manager [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 668.222072] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5c19632-32d0-4ab6-a394-3adf39c9c4e3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.229753] env[62385]: DEBUG nova.compute.manager [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 668.233405] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ee83371-dd66-49ea-a6d1-65b023f31471 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.267889] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a93fb324-e769-4a00-9eed-fb4706c8b24a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.275998] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af72b512-7f6e-4ced-bb14-18dadfaccccd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.291388] env[62385]: DEBUG nova.compute.provider_tree [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 668.400686] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.494686] env[62385]: DEBUG nova.network.neutron [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Successfully created port: ca62abe6-ab3c-4b9f-ab5e-ad10b81e8be9 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 668.796341] env[62385]: DEBUG nova.scheduler.client.report [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 669.243487] env[62385]: DEBUG nova.compute.manager [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 669.272690] env[62385]: DEBUG nova.virt.hardware [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 669.272806] env[62385]: DEBUG nova.virt.hardware [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 669.272896] env[62385]: DEBUG nova.virt.hardware [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 669.273085] env[62385]: DEBUG nova.virt.hardware [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 669.273249] env[62385]: DEBUG nova.virt.hardware [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 669.273510] env[62385]: DEBUG nova.virt.hardware [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 669.273611] env[62385]: DEBUG nova.virt.hardware [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 669.274044] env[62385]: DEBUG nova.virt.hardware [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 669.274044] env[62385]: DEBUG nova.virt.hardware [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 669.274190] env[62385]: DEBUG nova.virt.hardware [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 669.274479] env[62385]: DEBUG nova.virt.hardware [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 669.275371] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cf46492-c6a7-4b0b-a07c-a56dd7338aa5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.283555] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52156475-cc25-4d1b-8680-95d4a53735b9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.302603] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.092s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 669.303259] env[62385]: ERROR nova.compute.manager [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 99d91aa3-3018-4ac1-b1c4-c3f41b5cf2f9, please check neutron logs for more information. [ 669.303259] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Traceback (most recent call last): [ 669.303259] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 669.303259] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] self.driver.spawn(context, instance, image_meta, [ 669.303259] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 669.303259] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] self._vmops.spawn(context, instance, image_meta, injected_files, [ 669.303259] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 669.303259] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] vm_ref = self.build_virtual_machine(instance, [ 669.303259] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 669.303259] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] vif_infos = vmwarevif.get_vif_info(self._session, [ 669.303259] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 669.303569] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] for vif in network_info: [ 669.303569] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 669.303569] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] return self._sync_wrapper(fn, *args, **kwargs) [ 669.303569] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 669.303569] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] self.wait() [ 669.303569] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 669.303569] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] self[:] = self._gt.wait() [ 669.303569] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 669.303569] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] return self._exit_event.wait() [ 669.303569] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 669.303569] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] result = hub.switch() [ 669.303569] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 669.303569] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] return self.greenlet.switch() [ 669.303869] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 669.303869] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] result = function(*args, **kwargs) [ 669.303869] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 669.303869] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] return func(*args, **kwargs) [ 669.303869] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 669.303869] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] raise e [ 669.303869] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.303869] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] nwinfo = self.network_api.allocate_for_instance( [ 669.303869] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 669.303869] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] created_port_ids = self._update_ports_for_instance( [ 669.303869] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 669.303869] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] with excutils.save_and_reraise_exception(): [ 669.303869] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.304253] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] self.force_reraise() [ 669.304253] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.304253] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] raise self.value [ 669.304253] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 669.304253] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] updated_port = self._update_port( [ 669.304253] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.304253] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] _ensure_no_port_binding_failure(port) [ 669.304253] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.304253] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] raise exception.PortBindingFailed(port_id=port['id']) [ 669.304253] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] nova.exception.PortBindingFailed: Binding failed for port 99d91aa3-3018-4ac1-b1c4-c3f41b5cf2f9, please check neutron logs for more information. [ 669.304253] env[62385]: ERROR nova.compute.manager [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] [ 669.304660] env[62385]: DEBUG nova.compute.utils [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Binding failed for port 99d91aa3-3018-4ac1-b1c4-c3f41b5cf2f9, please check neutron logs for more information. {{(pid=62385) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 669.305789] env[62385]: DEBUG nova.compute.manager [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Build of instance c348ef22-abb0-48cf-98a6-3a4b217dd186 was re-scheduled: Binding failed for port 99d91aa3-3018-4ac1-b1c4-c3f41b5cf2f9, please check neutron logs for more information. {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 669.306206] env[62385]: DEBUG nova.compute.manager [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Unplugging VIFs for instance {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 669.306432] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] Acquiring lock "refresh_cache-c348ef22-abb0-48cf-98a6-3a4b217dd186" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 669.306579] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] Acquired lock "refresh_cache-c348ef22-abb0-48cf-98a6-3a4b217dd186" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.306738] env[62385]: DEBUG nova.network.neutron [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 669.308490] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ae8492d9-8f89-48fe-aa8c-4b169b2b533c tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.243s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 669.308838] env[62385]: DEBUG nova.objects.instance [None req-ae8492d9-8f89-48fe-aa8c-4b169b2b533c tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Lazy-loading 'resources' on Instance uuid f0f0eed0-a76a-42bc-9cca-91b93d7ad64c {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 669.641689] env[62385]: DEBUG nova.compute.manager [req-804847f2-dac0-440a-b641-ad3f33f94f7f req-be4da2c2-b119-4f38-a646-f58dc1dec555 service nova] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Received event network-changed-ca62abe6-ab3c-4b9f-ab5e-ad10b81e8be9 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 669.641891] env[62385]: DEBUG nova.compute.manager [req-804847f2-dac0-440a-b641-ad3f33f94f7f req-be4da2c2-b119-4f38-a646-f58dc1dec555 service nova] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Refreshing instance network info cache due to event network-changed-ca62abe6-ab3c-4b9f-ab5e-ad10b81e8be9. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 669.642162] env[62385]: DEBUG oslo_concurrency.lockutils [req-804847f2-dac0-440a-b641-ad3f33f94f7f req-be4da2c2-b119-4f38-a646-f58dc1dec555 service nova] Acquiring lock "refresh_cache-6e0dd86b-d557-4e7a-8520-9ffbd291128b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 669.642323] env[62385]: DEBUG oslo_concurrency.lockutils [req-804847f2-dac0-440a-b641-ad3f33f94f7f req-be4da2c2-b119-4f38-a646-f58dc1dec555 service nova] Acquired lock "refresh_cache-6e0dd86b-d557-4e7a-8520-9ffbd291128b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.642534] env[62385]: DEBUG nova.network.neutron [req-804847f2-dac0-440a-b641-ad3f33f94f7f req-be4da2c2-b119-4f38-a646-f58dc1dec555 service nova] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Refreshing network info cache for port ca62abe6-ab3c-4b9f-ab5e-ad10b81e8be9 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 669.818753] env[62385]: ERROR nova.compute.manager [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ca62abe6-ab3c-4b9f-ab5e-ad10b81e8be9, please check neutron logs for more information. [ 669.818753] env[62385]: ERROR nova.compute.manager Traceback (most recent call last): [ 669.818753] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.818753] env[62385]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 669.818753] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 669.818753] env[62385]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 669.818753] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 669.818753] env[62385]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 669.818753] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.818753] env[62385]: ERROR nova.compute.manager self.force_reraise() [ 669.818753] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.818753] env[62385]: ERROR nova.compute.manager raise self.value [ 669.818753] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 669.818753] env[62385]: ERROR nova.compute.manager updated_port = self._update_port( [ 669.818753] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.818753] env[62385]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 669.819441] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.819441] env[62385]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 669.819441] env[62385]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ca62abe6-ab3c-4b9f-ab5e-ad10b81e8be9, please check neutron logs for more information. [ 669.819441] env[62385]: ERROR nova.compute.manager [ 669.819441] env[62385]: Traceback (most recent call last): [ 669.819441] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 669.819441] env[62385]: listener.cb(fileno) [ 669.819441] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 669.819441] env[62385]: result = function(*args, **kwargs) [ 669.819441] env[62385]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 669.819441] env[62385]: return func(*args, **kwargs) [ 669.819441] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 669.819441] env[62385]: raise e [ 669.819441] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.819441] env[62385]: nwinfo = self.network_api.allocate_for_instance( [ 669.819441] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 669.819441] env[62385]: created_port_ids = self._update_ports_for_instance( [ 669.819441] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 669.819441] env[62385]: with excutils.save_and_reraise_exception(): [ 669.819441] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.819441] env[62385]: self.force_reraise() [ 669.819441] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.819441] env[62385]: raise self.value [ 669.819441] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 669.819441] env[62385]: updated_port = self._update_port( [ 669.819441] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.819441] env[62385]: _ensure_no_port_binding_failure(port) [ 669.819441] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.819441] env[62385]: raise exception.PortBindingFailed(port_id=port['id']) [ 669.820277] env[62385]: nova.exception.PortBindingFailed: Binding failed for port ca62abe6-ab3c-4b9f-ab5e-ad10b81e8be9, please check neutron logs for more information. [ 669.820277] env[62385]: Removing descriptor: 19 [ 669.820277] env[62385]: ERROR nova.compute.manager [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ca62abe6-ab3c-4b9f-ab5e-ad10b81e8be9, please check neutron logs for more information. [ 669.820277] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Traceback (most recent call last): [ 669.820277] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 669.820277] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] yield resources [ 669.820277] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 669.820277] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] self.driver.spawn(context, instance, image_meta, [ 669.820277] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 669.820277] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 669.820277] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 669.820277] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] vm_ref = self.build_virtual_machine(instance, [ 669.820739] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 669.820739] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] vif_infos = vmwarevif.get_vif_info(self._session, [ 669.820739] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 669.820739] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] for vif in network_info: [ 669.820739] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 669.820739] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] return self._sync_wrapper(fn, *args, **kwargs) [ 669.820739] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 669.820739] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] self.wait() [ 669.820739] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 669.820739] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] self[:] = self._gt.wait() [ 669.820739] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 669.820739] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] return self._exit_event.wait() [ 669.820739] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 669.824315] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] result = hub.switch() [ 669.824315] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 669.824315] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] return self.greenlet.switch() [ 669.824315] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 669.824315] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] result = function(*args, **kwargs) [ 669.824315] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 669.824315] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] return func(*args, **kwargs) [ 669.824315] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 669.824315] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] raise e [ 669.824315] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.824315] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] nwinfo = self.network_api.allocate_for_instance( [ 669.824315] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 669.824315] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] created_port_ids = self._update_ports_for_instance( [ 669.824702] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 669.824702] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] with excutils.save_and_reraise_exception(): [ 669.824702] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.824702] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] self.force_reraise() [ 669.824702] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.824702] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] raise self.value [ 669.824702] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 669.824702] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] updated_port = self._update_port( [ 669.824702] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.824702] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] _ensure_no_port_binding_failure(port) [ 669.824702] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.824702] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] raise exception.PortBindingFailed(port_id=port['id']) [ 669.825048] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] nova.exception.PortBindingFailed: Binding failed for port ca62abe6-ab3c-4b9f-ab5e-ad10b81e8be9, please check neutron logs for more information. [ 669.825048] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] [ 669.825048] env[62385]: INFO nova.compute.manager [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Terminating instance [ 669.826509] env[62385]: DEBUG oslo_concurrency.lockutils [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Acquiring lock "refresh_cache-6e0dd86b-d557-4e7a-8520-9ffbd291128b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 669.831625] env[62385]: DEBUG nova.network.neutron [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 669.964893] env[62385]: DEBUG nova.network.neutron [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.168716] env[62385]: DEBUG nova.network.neutron [req-804847f2-dac0-440a-b641-ad3f33f94f7f req-be4da2c2-b119-4f38-a646-f58dc1dec555 service nova] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 670.265098] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea228f04-6d08-4a64-b0d0-107a9552d24a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.271904] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63b1e1d1-1410-4c34-a053-54f2a8e2b2d6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.302688] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68e7aa60-011a-4391-8d43-33d492ed0a9f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.310688] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eb7f7cd-4f1a-4738-9628-44510aa43e26 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.325396] env[62385]: DEBUG nova.compute.provider_tree [None req-ae8492d9-8f89-48fe-aa8c-4b169b2b533c tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 670.327826] env[62385]: DEBUG nova.network.neutron [req-804847f2-dac0-440a-b641-ad3f33f94f7f req-be4da2c2-b119-4f38-a646-f58dc1dec555 service nova] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.469856] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] Releasing lock "refresh_cache-c348ef22-abb0-48cf-98a6-3a4b217dd186" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 670.470647] env[62385]: DEBUG nova.compute.manager [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 670.470647] env[62385]: DEBUG nova.compute.manager [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 670.470647] env[62385]: DEBUG nova.network.neutron [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 670.498886] env[62385]: DEBUG nova.network.neutron [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 670.831327] env[62385]: DEBUG nova.scheduler.client.report [None req-ae8492d9-8f89-48fe-aa8c-4b169b2b533c tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 670.837760] env[62385]: DEBUG oslo_concurrency.lockutils [req-804847f2-dac0-440a-b641-ad3f33f94f7f req-be4da2c2-b119-4f38-a646-f58dc1dec555 service nova] Releasing lock "refresh_cache-6e0dd86b-d557-4e7a-8520-9ffbd291128b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 670.837760] env[62385]: DEBUG oslo_concurrency.lockutils [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Acquired lock "refresh_cache-6e0dd86b-d557-4e7a-8520-9ffbd291128b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.837760] env[62385]: DEBUG nova.network.neutron [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 671.003491] env[62385]: DEBUG nova.network.neutron [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.339259] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ae8492d9-8f89-48fe-aa8c-4b169b2b533c tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.031s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 671.345331] env[62385]: DEBUG oslo_concurrency.lockutils [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.175s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 671.346636] env[62385]: INFO nova.compute.claims [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 671.364647] env[62385]: INFO nova.scheduler.client.report [None req-ae8492d9-8f89-48fe-aa8c-4b169b2b533c tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Deleted allocations for instance f0f0eed0-a76a-42bc-9cca-91b93d7ad64c [ 671.369117] env[62385]: DEBUG nova.network.neutron [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 671.489780] env[62385]: DEBUG nova.network.neutron [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.505743] env[62385]: INFO nova.compute.manager [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] [instance: c348ef22-abb0-48cf-98a6-3a4b217dd186] Took 1.04 seconds to deallocate network for instance. [ 671.813577] env[62385]: DEBUG nova.compute.manager [req-224ffb68-3bf3-4544-9338-0a16d7882909 req-2f609e2e-166f-4e3c-8d46-a14bcb66b73e service nova] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Received event network-vif-deleted-ca62abe6-ab3c-4b9f-ab5e-ad10b81e8be9 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 671.884237] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ae8492d9-8f89-48fe-aa8c-4b169b2b533c tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Lock "f0f0eed0-a76a-42bc-9cca-91b93d7ad64c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.542s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 671.992282] env[62385]: DEBUG oslo_concurrency.lockutils [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Releasing lock "refresh_cache-6e0dd86b-d557-4e7a-8520-9ffbd291128b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 671.992720] env[62385]: DEBUG nova.compute.manager [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 671.992911] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 671.993253] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a89883ee-17db-490f-8f32-6d4a108f7148 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.004430] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2597eadb-bbf5-4adc-9695-ad77172d8990 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.034531] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6e0dd86b-d557-4e7a-8520-9ffbd291128b could not be found. [ 672.034765] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 672.034951] env[62385]: INFO nova.compute.manager [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 672.035219] env[62385]: DEBUG oslo.service.loopingcall [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 672.035447] env[62385]: DEBUG nova.compute.manager [-] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 672.035540] env[62385]: DEBUG nova.network.neutron [-] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 672.051650] env[62385]: DEBUG nova.network.neutron [-] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 672.544314] env[62385]: INFO nova.scheduler.client.report [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] Deleted allocations for instance c348ef22-abb0-48cf-98a6-3a4b217dd186 [ 672.556632] env[62385]: DEBUG nova.network.neutron [-] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.780166] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c9d2357-b7c8-4dee-a2aa-77109d9b6989 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.790222] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ca452e7-6eb2-4fe2-9fed-8b8e47f7cf2d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.829553] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a190495-1580-4400-b0fd-c835b0bfd75a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.832412] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Acquiring lock "f77dc5ab-7e63-4a1b-9c87-e474a7e28c93" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 672.832666] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Lock "f77dc5ab-7e63-4a1b-9c87-e474a7e28c93" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 672.838227] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92eb625d-ebb1-4db6-9727-416a16a377e4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.851758] env[62385]: DEBUG nova.compute.provider_tree [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 672.941152] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Acquiring lock "8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 672.941420] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Lock "8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 673.016570] env[62385]: DEBUG oslo_concurrency.lockutils [None req-50e22468-1913-42eb-9f13-0da6a2f22848 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Acquiring lock "0ba32627-c646-477d-9062-b84624d01e48" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.016836] env[62385]: DEBUG oslo_concurrency.lockutils [None req-50e22468-1913-42eb-9f13-0da6a2f22848 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Lock "0ba32627-c646-477d-9062-b84624d01e48" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 673.017144] env[62385]: DEBUG oslo_concurrency.lockutils [None req-50e22468-1913-42eb-9f13-0da6a2f22848 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Acquiring lock "0ba32627-c646-477d-9062-b84624d01e48-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.017349] env[62385]: DEBUG oslo_concurrency.lockutils [None req-50e22468-1913-42eb-9f13-0da6a2f22848 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Lock "0ba32627-c646-477d-9062-b84624d01e48-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 673.017522] env[62385]: DEBUG oslo_concurrency.lockutils [None req-50e22468-1913-42eb-9f13-0da6a2f22848 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Lock "0ba32627-c646-477d-9062-b84624d01e48-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.019728] env[62385]: INFO nova.compute.manager [None req-50e22468-1913-42eb-9f13-0da6a2f22848 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Terminating instance [ 673.022554] env[62385]: DEBUG oslo_concurrency.lockutils [None req-50e22468-1913-42eb-9f13-0da6a2f22848 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Acquiring lock "refresh_cache-0ba32627-c646-477d-9062-b84624d01e48" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 673.023022] env[62385]: DEBUG oslo_concurrency.lockutils [None req-50e22468-1913-42eb-9f13-0da6a2f22848 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Acquired lock "refresh_cache-0ba32627-c646-477d-9062-b84624d01e48" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 673.023246] env[62385]: DEBUG nova.network.neutron [None req-50e22468-1913-42eb-9f13-0da6a2f22848 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 673.055135] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6302dd3a-4ef3-4cfc-ab1d-466be490a8e9 tempest-ServerActionsTestOtherB-1036312694 tempest-ServerActionsTestOtherB-1036312694-project-member] Lock "c348ef22-abb0-48cf-98a6-3a4b217dd186" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.545s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.059134] env[62385]: INFO nova.compute.manager [-] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Took 1.02 seconds to deallocate network for instance. [ 673.061584] env[62385]: DEBUG nova.compute.claims [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Aborting claim: {{(pid=62385) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 673.061584] env[62385]: DEBUG oslo_concurrency.lockutils [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.355028] env[62385]: DEBUG nova.scheduler.client.report [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 673.545926] env[62385]: DEBUG nova.network.neutron [None req-50e22468-1913-42eb-9f13-0da6a2f22848 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 673.557491] env[62385]: DEBUG nova.compute.manager [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 673.636085] env[62385]: DEBUG nova.network.neutron [None req-50e22468-1913-42eb-9f13-0da6a2f22848 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.860487] env[62385]: DEBUG oslo_concurrency.lockutils [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.515s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.861033] env[62385]: DEBUG nova.compute.manager [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 673.863562] env[62385]: DEBUG oslo_concurrency.lockutils [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 32.636s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.079545] env[62385]: DEBUG oslo_concurrency.lockutils [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.139378] env[62385]: DEBUG oslo_concurrency.lockutils [None req-50e22468-1913-42eb-9f13-0da6a2f22848 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Releasing lock "refresh_cache-0ba32627-c646-477d-9062-b84624d01e48" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 674.139790] env[62385]: DEBUG nova.compute.manager [None req-50e22468-1913-42eb-9f13-0da6a2f22848 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 674.139990] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-50e22468-1913-42eb-9f13-0da6a2f22848 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 674.140890] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6036c720-cb5a-40e5-930f-158f68960038 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.148643] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-50e22468-1913-42eb-9f13-0da6a2f22848 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 674.148869] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cbe5d9d9-d350-4085-bef7-bc80f5374e58 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.156057] env[62385]: DEBUG oslo_vmware.api [None req-50e22468-1913-42eb-9f13-0da6a2f22848 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Waiting for the task: (returnval){ [ 674.156057] env[62385]: value = "task-1205479" [ 674.156057] env[62385]: _type = "Task" [ 674.156057] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.163839] env[62385]: DEBUG oslo_vmware.api [None req-50e22468-1913-42eb-9f13-0da6a2f22848 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205479, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.370034] env[62385]: DEBUG nova.compute.utils [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 674.375217] env[62385]: DEBUG nova.compute.manager [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 674.375366] env[62385]: DEBUG nova.network.neutron [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 674.427929] env[62385]: DEBUG nova.policy [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4f9a152126e44c72846fbf187e17e820', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ddb3ad4762cc4027a207fcab3a5eed5b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 674.667082] env[62385]: DEBUG oslo_vmware.api [None req-50e22468-1913-42eb-9f13-0da6a2f22848 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205479, 'name': PowerOffVM_Task, 'duration_secs': 0.113546} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.671277] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-50e22468-1913-42eb-9f13-0da6a2f22848 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 674.671277] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-50e22468-1913-42eb-9f13-0da6a2f22848 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 674.674809] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-27e6f3b8-f87e-4032-9749-65bbccfe3133 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.704503] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-50e22468-1913-42eb-9f13-0da6a2f22848 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 674.704701] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-50e22468-1913-42eb-9f13-0da6a2f22848 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Deleting contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 674.704876] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-50e22468-1913-42eb-9f13-0da6a2f22848 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Deleting the datastore file [datastore2] 0ba32627-c646-477d-9062-b84624d01e48 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 674.705698] env[62385]: DEBUG nova.network.neutron [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Successfully created port: cb327ab2-1ebf-4a2a-9aec-b41148bf5183 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 674.709202] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3cb465f4-2625-4784-b95c-3c71005424ab {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.718080] env[62385]: DEBUG oslo_vmware.api [None req-50e22468-1913-42eb-9f13-0da6a2f22848 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Waiting for the task: (returnval){ [ 674.718080] env[62385]: value = "task-1205481" [ 674.718080] env[62385]: _type = "Task" [ 674.718080] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.725989] env[62385]: DEBUG oslo_vmware.api [None req-50e22468-1913-42eb-9f13-0da6a2f22848 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205481, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.826741] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb948a6f-bdeb-49c7-8b38-099e36b06adb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.833519] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a84a63bd-44ab-4f9c-b8e3-a6bdc550ac96 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.871610] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1153e7b8-aa63-47bc-93e2-6d557b1e9e04 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.877041] env[62385]: DEBUG nova.compute.manager [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 674.882808] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0101067-d50f-4a1e-ad41-0126d9c09dc2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.898950] env[62385]: DEBUG nova.compute.provider_tree [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 675.227587] env[62385]: DEBUG oslo_vmware.api [None req-50e22468-1913-42eb-9f13-0da6a2f22848 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Task: {'id': task-1205481, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.102284} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.227845] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-50e22468-1913-42eb-9f13-0da6a2f22848 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 675.228267] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-50e22468-1913-42eb-9f13-0da6a2f22848 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Deleted contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 675.228267] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-50e22468-1913-42eb-9f13-0da6a2f22848 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 675.228436] env[62385]: INFO nova.compute.manager [None req-50e22468-1913-42eb-9f13-0da6a2f22848 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Took 1.09 seconds to destroy the instance on the hypervisor. [ 675.228670] env[62385]: DEBUG oslo.service.loopingcall [None req-50e22468-1913-42eb-9f13-0da6a2f22848 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 675.228854] env[62385]: DEBUG nova.compute.manager [-] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 675.228948] env[62385]: DEBUG nova.network.neutron [-] [instance: 0ba32627-c646-477d-9062-b84624d01e48] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 675.246912] env[62385]: DEBUG nova.network.neutron [-] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 675.403776] env[62385]: DEBUG nova.scheduler.client.report [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 675.424402] env[62385]: DEBUG nova.compute.manager [req-b85d5063-f5fd-4bfb-af72-013f0f2b2029 req-3453aa65-3ce8-432e-be80-f1f477386ec2 service nova] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Received event network-changed-cb327ab2-1ebf-4a2a-9aec-b41148bf5183 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 675.425534] env[62385]: DEBUG nova.compute.manager [req-b85d5063-f5fd-4bfb-af72-013f0f2b2029 req-3453aa65-3ce8-432e-be80-f1f477386ec2 service nova] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Refreshing instance network info cache due to event network-changed-cb327ab2-1ebf-4a2a-9aec-b41148bf5183. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 675.425773] env[62385]: DEBUG oslo_concurrency.lockutils [req-b85d5063-f5fd-4bfb-af72-013f0f2b2029 req-3453aa65-3ce8-432e-be80-f1f477386ec2 service nova] Acquiring lock "refresh_cache-ea0b59fe-8dd9-4893-87ff-54706eb64479" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 675.425989] env[62385]: DEBUG oslo_concurrency.lockutils [req-b85d5063-f5fd-4bfb-af72-013f0f2b2029 req-3453aa65-3ce8-432e-be80-f1f477386ec2 service nova] Acquired lock "refresh_cache-ea0b59fe-8dd9-4893-87ff-54706eb64479" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.426229] env[62385]: DEBUG nova.network.neutron [req-b85d5063-f5fd-4bfb-af72-013f0f2b2029 req-3453aa65-3ce8-432e-be80-f1f477386ec2 service nova] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Refreshing network info cache for port cb327ab2-1ebf-4a2a-9aec-b41148bf5183 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 675.614982] env[62385]: ERROR nova.compute.manager [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port cb327ab2-1ebf-4a2a-9aec-b41148bf5183, please check neutron logs for more information. [ 675.614982] env[62385]: ERROR nova.compute.manager Traceback (most recent call last): [ 675.614982] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 675.614982] env[62385]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 675.614982] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 675.614982] env[62385]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 675.614982] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 675.614982] env[62385]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 675.614982] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.614982] env[62385]: ERROR nova.compute.manager self.force_reraise() [ 675.614982] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.614982] env[62385]: ERROR nova.compute.manager raise self.value [ 675.614982] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 675.614982] env[62385]: ERROR nova.compute.manager updated_port = self._update_port( [ 675.614982] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.614982] env[62385]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 675.615474] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.615474] env[62385]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 675.615474] env[62385]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port cb327ab2-1ebf-4a2a-9aec-b41148bf5183, please check neutron logs for more information. [ 675.615474] env[62385]: ERROR nova.compute.manager [ 675.615474] env[62385]: Traceback (most recent call last): [ 675.615616] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 675.615616] env[62385]: listener.cb(fileno) [ 675.615616] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 675.615616] env[62385]: result = function(*args, **kwargs) [ 675.615616] env[62385]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 675.615616] env[62385]: return func(*args, **kwargs) [ 675.615616] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 675.615616] env[62385]: raise e [ 675.615616] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 675.615616] env[62385]: nwinfo = self.network_api.allocate_for_instance( [ 675.615616] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 675.615616] env[62385]: created_port_ids = self._update_ports_for_instance( [ 675.615616] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 675.615616] env[62385]: with excutils.save_and_reraise_exception(): [ 675.615616] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.615616] env[62385]: self.force_reraise() [ 675.615616] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.615616] env[62385]: raise self.value [ 675.615616] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 675.615616] env[62385]: updated_port = self._update_port( [ 675.615616] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.615616] env[62385]: _ensure_no_port_binding_failure(port) [ 675.615616] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.615616] env[62385]: raise exception.PortBindingFailed(port_id=port['id']) [ 675.615616] env[62385]: nova.exception.PortBindingFailed: Binding failed for port cb327ab2-1ebf-4a2a-9aec-b41148bf5183, please check neutron logs for more information. [ 675.615616] env[62385]: Removing descriptor: 17 [ 675.750178] env[62385]: DEBUG nova.network.neutron [-] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 675.888055] env[62385]: DEBUG nova.compute.manager [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 675.908763] env[62385]: DEBUG oslo_concurrency.lockutils [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.045s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 675.909392] env[62385]: ERROR nova.compute.manager [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] [instance: ad12e5b0-6854-4281-9155-e1660393829c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port be90e400-b8be-4151-9773-122efa78f030, please check neutron logs for more information. [ 675.909392] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] Traceback (most recent call last): [ 675.909392] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 675.909392] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] self.driver.spawn(context, instance, image_meta, [ 675.909392] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 675.909392] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 675.909392] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 675.909392] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] vm_ref = self.build_virtual_machine(instance, [ 675.909392] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 675.909392] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] vif_infos = vmwarevif.get_vif_info(self._session, [ 675.909392] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 675.909719] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] for vif in network_info: [ 675.909719] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 675.909719] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] return self._sync_wrapper(fn, *args, **kwargs) [ 675.909719] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 675.909719] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] self.wait() [ 675.909719] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 675.909719] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] self[:] = self._gt.wait() [ 675.909719] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 675.909719] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] return self._exit_event.wait() [ 675.909719] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 675.909719] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] result = hub.switch() [ 675.909719] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 675.909719] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] return self.greenlet.switch() [ 675.910049] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 675.910049] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] result = function(*args, **kwargs) [ 675.910049] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 675.910049] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] return func(*args, **kwargs) [ 675.910049] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 675.910049] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] raise e [ 675.910049] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 675.910049] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] nwinfo = self.network_api.allocate_for_instance( [ 675.910049] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 675.910049] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] created_port_ids = self._update_ports_for_instance( [ 675.910049] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 675.910049] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] with excutils.save_and_reraise_exception(): [ 675.910049] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.910377] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] self.force_reraise() [ 675.910377] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.910377] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] raise self.value [ 675.910377] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 675.910377] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] updated_port = self._update_port( [ 675.910377] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.910377] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] _ensure_no_port_binding_failure(port) [ 675.910377] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.910377] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] raise exception.PortBindingFailed(port_id=port['id']) [ 675.910377] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] nova.exception.PortBindingFailed: Binding failed for port be90e400-b8be-4151-9773-122efa78f030, please check neutron logs for more information. [ 675.910377] env[62385]: ERROR nova.compute.manager [instance: ad12e5b0-6854-4281-9155-e1660393829c] [ 675.910652] env[62385]: DEBUG nova.compute.utils [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] [instance: ad12e5b0-6854-4281-9155-e1660393829c] Binding failed for port be90e400-b8be-4151-9773-122efa78f030, please check neutron logs for more information. {{(pid=62385) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 675.913936] env[62385]: DEBUG nova.virt.hardware [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 675.914174] env[62385]: DEBUG nova.virt.hardware [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 675.914334] env[62385]: DEBUG nova.virt.hardware [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 675.914517] env[62385]: DEBUG nova.virt.hardware [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 675.914664] env[62385]: DEBUG nova.virt.hardware [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 675.914812] env[62385]: DEBUG nova.virt.hardware [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 675.915041] env[62385]: DEBUG nova.virt.hardware [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 675.915214] env[62385]: DEBUG nova.virt.hardware [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 675.915385] env[62385]: DEBUG nova.virt.hardware [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 675.915547] env[62385]: DEBUG nova.virt.hardware [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 675.915715] env[62385]: DEBUG nova.virt.hardware [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 675.916034] env[62385]: DEBUG oslo_concurrency.lockutils [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.684s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 675.917490] env[62385]: INFO nova.compute.claims [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 675.921262] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea83d9bf-5a83-447d-a545-fc54de2a14d6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.924363] env[62385]: DEBUG nova.compute.manager [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] [instance: ad12e5b0-6854-4281-9155-e1660393829c] Build of instance ad12e5b0-6854-4281-9155-e1660393829c was re-scheduled: Binding failed for port be90e400-b8be-4151-9773-122efa78f030, please check neutron logs for more information. {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 675.924817] env[62385]: DEBUG nova.compute.manager [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] [instance: ad12e5b0-6854-4281-9155-e1660393829c] Unplugging VIFs for instance {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 675.925053] env[62385]: DEBUG oslo_concurrency.lockutils [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] Acquiring lock "refresh_cache-ad12e5b0-6854-4281-9155-e1660393829c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 675.925669] env[62385]: DEBUG oslo_concurrency.lockutils [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] Acquired lock "refresh_cache-ad12e5b0-6854-4281-9155-e1660393829c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.925669] env[62385]: DEBUG nova.network.neutron [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] [instance: ad12e5b0-6854-4281-9155-e1660393829c] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 675.934563] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02144487-8eaf-42f1-aee6-806b4a3e7227 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.949856] env[62385]: ERROR nova.compute.manager [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port cb327ab2-1ebf-4a2a-9aec-b41148bf5183, please check neutron logs for more information. [ 675.949856] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Traceback (most recent call last): [ 675.949856] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 675.949856] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] yield resources [ 675.949856] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 675.949856] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] self.driver.spawn(context, instance, image_meta, [ 675.949856] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 675.949856] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] self._vmops.spawn(context, instance, image_meta, injected_files, [ 675.949856] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 675.949856] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] vm_ref = self.build_virtual_machine(instance, [ 675.949856] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 675.950306] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] vif_infos = vmwarevif.get_vif_info(self._session, [ 675.950306] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 675.950306] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] for vif in network_info: [ 675.950306] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 675.950306] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] return self._sync_wrapper(fn, *args, **kwargs) [ 675.950306] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 675.950306] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] self.wait() [ 675.950306] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 675.950306] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] self[:] = self._gt.wait() [ 675.950306] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 675.950306] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] return self._exit_event.wait() [ 675.950306] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 675.950306] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] current.throw(*self._exc) [ 675.950692] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 675.950692] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] result = function(*args, **kwargs) [ 675.950692] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 675.950692] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] return func(*args, **kwargs) [ 675.950692] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 675.950692] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] raise e [ 675.950692] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 675.950692] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] nwinfo = self.network_api.allocate_for_instance( [ 675.950692] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 675.950692] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] created_port_ids = self._update_ports_for_instance( [ 675.950692] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 675.950692] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] with excutils.save_and_reraise_exception(): [ 675.950692] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.951088] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] self.force_reraise() [ 675.951088] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.951088] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] raise self.value [ 675.951088] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 675.951088] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] updated_port = self._update_port( [ 675.951088] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.951088] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] _ensure_no_port_binding_failure(port) [ 675.951088] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.951088] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] raise exception.PortBindingFailed(port_id=port['id']) [ 675.951088] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] nova.exception.PortBindingFailed: Binding failed for port cb327ab2-1ebf-4a2a-9aec-b41148bf5183, please check neutron logs for more information. [ 675.951088] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] [ 675.951088] env[62385]: INFO nova.compute.manager [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Terminating instance [ 675.952180] env[62385]: DEBUG oslo_concurrency.lockutils [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] Acquiring lock "refresh_cache-ea0b59fe-8dd9-4893-87ff-54706eb64479" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 675.954465] env[62385]: DEBUG nova.network.neutron [req-b85d5063-f5fd-4bfb-af72-013f0f2b2029 req-3453aa65-3ce8-432e-be80-f1f477386ec2 service nova] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 676.036061] env[62385]: DEBUG nova.network.neutron [req-b85d5063-f5fd-4bfb-af72-013f0f2b2029 req-3453aa65-3ce8-432e-be80-f1f477386ec2 service nova] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.253565] env[62385]: INFO nova.compute.manager [-] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Took 1.02 seconds to deallocate network for instance. [ 676.448980] env[62385]: DEBUG nova.network.neutron [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] [instance: ad12e5b0-6854-4281-9155-e1660393829c] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 676.499643] env[62385]: DEBUG nova.network.neutron [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] [instance: ad12e5b0-6854-4281-9155-e1660393829c] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.538854] env[62385]: DEBUG oslo_concurrency.lockutils [req-b85d5063-f5fd-4bfb-af72-013f0f2b2029 req-3453aa65-3ce8-432e-be80-f1f477386ec2 service nova] Releasing lock "refresh_cache-ea0b59fe-8dd9-4893-87ff-54706eb64479" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 676.539264] env[62385]: DEBUG oslo_concurrency.lockutils [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] Acquired lock "refresh_cache-ea0b59fe-8dd9-4893-87ff-54706eb64479" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.539451] env[62385]: DEBUG nova.network.neutron [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 676.760129] env[62385]: DEBUG oslo_concurrency.lockutils [None req-50e22468-1913-42eb-9f13-0da6a2f22848 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 677.002568] env[62385]: DEBUG oslo_concurrency.lockutils [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] Releasing lock "refresh_cache-ad12e5b0-6854-4281-9155-e1660393829c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 677.002813] env[62385]: DEBUG nova.compute.manager [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 677.002972] env[62385]: DEBUG nova.compute.manager [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] [instance: ad12e5b0-6854-4281-9155-e1660393829c] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 677.003151] env[62385]: DEBUG nova.network.neutron [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] [instance: ad12e5b0-6854-4281-9155-e1660393829c] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 677.019785] env[62385]: DEBUG nova.network.neutron [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] [instance: ad12e5b0-6854-4281-9155-e1660393829c] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 677.056685] env[62385]: DEBUG nova.network.neutron [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 677.131713] env[62385]: DEBUG nova.network.neutron [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.263913] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cf113f1-119b-478e-948d-17dad32c31f4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.271498] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47273ae0-e123-4f53-913d-fe247ded1c76 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.300525] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f092259-0d7e-41db-abee-7f169b34ac33 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.307565] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb5f5f68-5172-401e-a2ed-fdb106694c79 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.321296] env[62385]: DEBUG nova.compute.provider_tree [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 677.454556] env[62385]: DEBUG nova.compute.manager [req-2141ed88-267d-4812-a39b-051e538e2061 req-55f44b38-1cd1-4a35-827a-67c0c7e3491b service nova] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Received event network-vif-deleted-cb327ab2-1ebf-4a2a-9aec-b41148bf5183 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 677.523721] env[62385]: DEBUG nova.network.neutron [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] [instance: ad12e5b0-6854-4281-9155-e1660393829c] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.635666] env[62385]: DEBUG oslo_concurrency.lockutils [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] Releasing lock "refresh_cache-ea0b59fe-8dd9-4893-87ff-54706eb64479" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 677.636143] env[62385]: DEBUG nova.compute.manager [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 677.636354] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 677.636661] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-83dc1e8d-83b3-400b-93e1-23cdcf5aec59 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.645490] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c2c6158-c2b1-4b68-8295-4dc389ff3f4a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.666164] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ea0b59fe-8dd9-4893-87ff-54706eb64479 could not be found. [ 677.666461] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 677.666667] env[62385]: INFO nova.compute.manager [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Took 0.03 seconds to destroy the instance on the hypervisor. [ 677.666911] env[62385]: DEBUG oslo.service.loopingcall [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 677.667161] env[62385]: DEBUG nova.compute.manager [-] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 677.667276] env[62385]: DEBUG nova.network.neutron [-] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 677.681924] env[62385]: DEBUG nova.network.neutron [-] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 677.825629] env[62385]: DEBUG nova.scheduler.client.report [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 678.026545] env[62385]: INFO nova.compute.manager [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] [instance: ad12e5b0-6854-4281-9155-e1660393829c] Took 1.02 seconds to deallocate network for instance. [ 678.184818] env[62385]: DEBUG nova.network.neutron [-] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.331516] env[62385]: DEBUG oslo_concurrency.lockutils [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.415s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 678.332040] env[62385]: DEBUG nova.compute.manager [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 678.335016] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.261s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 678.688618] env[62385]: INFO nova.compute.manager [-] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Took 1.02 seconds to deallocate network for instance. [ 678.690476] env[62385]: DEBUG nova.compute.claims [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Aborting claim: {{(pid=62385) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 678.690676] env[62385]: DEBUG oslo_concurrency.lockutils [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 678.839312] env[62385]: DEBUG nova.compute.utils [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 678.844358] env[62385]: DEBUG nova.compute.manager [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 678.844699] env[62385]: DEBUG nova.network.neutron [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 678.894798] env[62385]: DEBUG nova.policy [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2260fe8cb079468a8683a6c8b0a1d90e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '44b653c0620f489bbbc4d0dfac21c8e1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 679.058693] env[62385]: INFO nova.scheduler.client.report [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] Deleted allocations for instance ad12e5b0-6854-4281-9155-e1660393829c [ 679.200280] env[62385]: DEBUG nova.network.neutron [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Successfully created port: 0776ce7e-e90a-4326-ac85-6cdc70f2e785 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 679.235251] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-957233fe-5570-4fe5-86c0-1cdecf72758d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.242774] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a402b9ce-ee27-45ac-a773-fc3d346c88f4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.275048] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50a3c933-66a8-4a11-88c7-5ade8f625917 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.280111] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-286d3a64-267b-4889-bc6e-53f86de89477 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.292767] env[62385]: DEBUG nova.compute.provider_tree [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 679.345221] env[62385]: DEBUG nova.compute.manager [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 679.567210] env[62385]: DEBUG oslo_concurrency.lockutils [None req-52a1e7a9-d88b-4b81-8a31-edd1306bb3c2 tempest-ServerExternalEventsTest-1484084025 tempest-ServerExternalEventsTest-1484084025-project-member] Lock "ad12e5b0-6854-4281-9155-e1660393829c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 124.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.798091] env[62385]: DEBUG nova.scheduler.client.report [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 679.836315] env[62385]: DEBUG nova.compute.manager [req-f90f899e-36a6-4251-a0a6-16e59fb7215f req-cb62ef0a-50c3-4b61-8232-7af9cf1f089c service nova] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Received event network-changed-0776ce7e-e90a-4326-ac85-6cdc70f2e785 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 679.836488] env[62385]: DEBUG nova.compute.manager [req-f90f899e-36a6-4251-a0a6-16e59fb7215f req-cb62ef0a-50c3-4b61-8232-7af9cf1f089c service nova] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Refreshing instance network info cache due to event network-changed-0776ce7e-e90a-4326-ac85-6cdc70f2e785. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 679.836668] env[62385]: DEBUG oslo_concurrency.lockutils [req-f90f899e-36a6-4251-a0a6-16e59fb7215f req-cb62ef0a-50c3-4b61-8232-7af9cf1f089c service nova] Acquiring lock "refresh_cache-7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 679.836809] env[62385]: DEBUG oslo_concurrency.lockutils [req-f90f899e-36a6-4251-a0a6-16e59fb7215f req-cb62ef0a-50c3-4b61-8232-7af9cf1f089c service nova] Acquired lock "refresh_cache-7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.836968] env[62385]: DEBUG nova.network.neutron [req-f90f899e-36a6-4251-a0a6-16e59fb7215f req-cb62ef0a-50c3-4b61-8232-7af9cf1f089c service nova] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Refreshing network info cache for port 0776ce7e-e90a-4326-ac85-6cdc70f2e785 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 680.032375] env[62385]: ERROR nova.compute.manager [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0776ce7e-e90a-4326-ac85-6cdc70f2e785, please check neutron logs for more information. [ 680.032375] env[62385]: ERROR nova.compute.manager Traceback (most recent call last): [ 680.032375] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 680.032375] env[62385]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 680.032375] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 680.032375] env[62385]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 680.032375] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 680.032375] env[62385]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 680.032375] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.032375] env[62385]: ERROR nova.compute.manager self.force_reraise() [ 680.032375] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.032375] env[62385]: ERROR nova.compute.manager raise self.value [ 680.032375] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 680.032375] env[62385]: ERROR nova.compute.manager updated_port = self._update_port( [ 680.032375] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.032375] env[62385]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 680.033043] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.033043] env[62385]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 680.033043] env[62385]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0776ce7e-e90a-4326-ac85-6cdc70f2e785, please check neutron logs for more information. [ 680.033043] env[62385]: ERROR nova.compute.manager [ 680.033043] env[62385]: Traceback (most recent call last): [ 680.033043] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 680.033043] env[62385]: listener.cb(fileno) [ 680.033043] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 680.033043] env[62385]: result = function(*args, **kwargs) [ 680.033043] env[62385]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 680.033043] env[62385]: return func(*args, **kwargs) [ 680.033043] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 680.033043] env[62385]: raise e [ 680.033043] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 680.033043] env[62385]: nwinfo = self.network_api.allocate_for_instance( [ 680.033043] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 680.033043] env[62385]: created_port_ids = self._update_ports_for_instance( [ 680.033043] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 680.033043] env[62385]: with excutils.save_and_reraise_exception(): [ 680.033043] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.033043] env[62385]: self.force_reraise() [ 680.033043] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.033043] env[62385]: raise self.value [ 680.033043] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 680.033043] env[62385]: updated_port = self._update_port( [ 680.033043] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.033043] env[62385]: _ensure_no_port_binding_failure(port) [ 680.033043] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.033043] env[62385]: raise exception.PortBindingFailed(port_id=port['id']) [ 680.033826] env[62385]: nova.exception.PortBindingFailed: Binding failed for port 0776ce7e-e90a-4326-ac85-6cdc70f2e785, please check neutron logs for more information. [ 680.033826] env[62385]: Removing descriptor: 17 [ 680.071020] env[62385]: DEBUG nova.compute.manager [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 680.303298] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.969s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.303930] env[62385]: ERROR nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: da94a435-b36c-42b2-8c84-564265872048] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 55745fbd-4c5f-410c-b72c-27ce94a14e33, please check neutron logs for more information. [ 680.303930] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] Traceback (most recent call last): [ 680.303930] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 680.303930] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] self.driver.spawn(context, instance, image_meta, [ 680.303930] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 680.303930] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] self._vmops.spawn(context, instance, image_meta, injected_files, [ 680.303930] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 680.303930] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] vm_ref = self.build_virtual_machine(instance, [ 680.303930] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 680.303930] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] vif_infos = vmwarevif.get_vif_info(self._session, [ 680.303930] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 680.304276] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] for vif in network_info: [ 680.304276] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 680.304276] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] return self._sync_wrapper(fn, *args, **kwargs) [ 680.304276] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 680.304276] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] self.wait() [ 680.304276] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 680.304276] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] self[:] = self._gt.wait() [ 680.304276] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 680.304276] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] return self._exit_event.wait() [ 680.304276] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 680.304276] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] result = hub.switch() [ 680.304276] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 680.304276] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] return self.greenlet.switch() [ 680.304649] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 680.304649] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] result = function(*args, **kwargs) [ 680.304649] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 680.304649] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] return func(*args, **kwargs) [ 680.304649] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 680.304649] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] raise e [ 680.304649] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 680.304649] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] nwinfo = self.network_api.allocate_for_instance( [ 680.304649] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 680.304649] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] created_port_ids = self._update_ports_for_instance( [ 680.304649] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 680.304649] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] with excutils.save_and_reraise_exception(): [ 680.304649] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.305045] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] self.force_reraise() [ 680.305045] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.305045] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] raise self.value [ 680.305045] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 680.305045] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] updated_port = self._update_port( [ 680.305045] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.305045] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] _ensure_no_port_binding_failure(port) [ 680.305045] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.305045] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] raise exception.PortBindingFailed(port_id=port['id']) [ 680.305045] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] nova.exception.PortBindingFailed: Binding failed for port 55745fbd-4c5f-410c-b72c-27ce94a14e33, please check neutron logs for more information. [ 680.305045] env[62385]: ERROR nova.compute.manager [instance: da94a435-b36c-42b2-8c84-564265872048] [ 680.305530] env[62385]: DEBUG nova.compute.utils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: da94a435-b36c-42b2-8c84-564265872048] Binding failed for port 55745fbd-4c5f-410c-b72c-27ce94a14e33, please check neutron logs for more information. {{(pid=62385) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 680.305939] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.698s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.309304] env[62385]: DEBUG nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: da94a435-b36c-42b2-8c84-564265872048] Build of instance da94a435-b36c-42b2-8c84-564265872048 was re-scheduled: Binding failed for port 55745fbd-4c5f-410c-b72c-27ce94a14e33, please check neutron logs for more information. {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 680.309744] env[62385]: DEBUG nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: da94a435-b36c-42b2-8c84-564265872048] Unplugging VIFs for instance {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 680.309963] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Acquiring lock "refresh_cache-da94a435-b36c-42b2-8c84-564265872048" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.310121] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Acquired lock "refresh_cache-da94a435-b36c-42b2-8c84-564265872048" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.310285] env[62385]: DEBUG nova.network.neutron [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: da94a435-b36c-42b2-8c84-564265872048] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 680.354450] env[62385]: DEBUG nova.compute.manager [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 680.357270] env[62385]: DEBUG nova.network.neutron [req-f90f899e-36a6-4251-a0a6-16e59fb7215f req-cb62ef0a-50c3-4b61-8232-7af9cf1f089c service nova] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 680.382384] env[62385]: DEBUG nova.virt.hardware [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 680.382674] env[62385]: DEBUG nova.virt.hardware [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 680.382887] env[62385]: DEBUG nova.virt.hardware [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 680.383013] env[62385]: DEBUG nova.virt.hardware [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 680.383296] env[62385]: DEBUG nova.virt.hardware [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 680.383451] env[62385]: DEBUG nova.virt.hardware [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 680.383665] env[62385]: DEBUG nova.virt.hardware [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 680.383877] env[62385]: DEBUG nova.virt.hardware [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 680.384082] env[62385]: DEBUG nova.virt.hardware [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 680.384252] env[62385]: DEBUG nova.virt.hardware [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 680.384442] env[62385]: DEBUG nova.virt.hardware [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 680.385390] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec954dce-f566-44a1-9c57-3ccc4ca2e37b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.393850] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b50ee971-8df4-45d9-8624-67b7ac9be8c2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.407776] env[62385]: ERROR nova.compute.manager [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0776ce7e-e90a-4326-ac85-6cdc70f2e785, please check neutron logs for more information. [ 680.407776] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Traceback (most recent call last): [ 680.407776] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 680.407776] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] yield resources [ 680.407776] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 680.407776] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] self.driver.spawn(context, instance, image_meta, [ 680.407776] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 680.407776] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] self._vmops.spawn(context, instance, image_meta, injected_files, [ 680.407776] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 680.407776] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] vm_ref = self.build_virtual_machine(instance, [ 680.407776] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 680.408207] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] vif_infos = vmwarevif.get_vif_info(self._session, [ 680.408207] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 680.408207] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] for vif in network_info: [ 680.408207] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 680.408207] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] return self._sync_wrapper(fn, *args, **kwargs) [ 680.408207] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 680.408207] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] self.wait() [ 680.408207] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 680.408207] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] self[:] = self._gt.wait() [ 680.408207] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 680.408207] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] return self._exit_event.wait() [ 680.408207] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 680.408207] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] current.throw(*self._exc) [ 680.408624] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 680.408624] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] result = function(*args, **kwargs) [ 680.408624] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 680.408624] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] return func(*args, **kwargs) [ 680.408624] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 680.408624] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] raise e [ 680.408624] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 680.408624] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] nwinfo = self.network_api.allocate_for_instance( [ 680.408624] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 680.408624] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] created_port_ids = self._update_ports_for_instance( [ 680.408624] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 680.408624] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] with excutils.save_and_reraise_exception(): [ 680.408624] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.409013] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] self.force_reraise() [ 680.409013] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.409013] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] raise self.value [ 680.409013] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 680.409013] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] updated_port = self._update_port( [ 680.409013] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.409013] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] _ensure_no_port_binding_failure(port) [ 680.409013] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.409013] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] raise exception.PortBindingFailed(port_id=port['id']) [ 680.409013] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] nova.exception.PortBindingFailed: Binding failed for port 0776ce7e-e90a-4326-ac85-6cdc70f2e785, please check neutron logs for more information. [ 680.409013] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] [ 680.409013] env[62385]: INFO nova.compute.manager [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Terminating instance [ 680.412212] env[62385]: DEBUG oslo_concurrency.lockutils [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] Acquiring lock "refresh_cache-7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.445759] env[62385]: DEBUG nova.network.neutron [req-f90f899e-36a6-4251-a0a6-16e59fb7215f req-cb62ef0a-50c3-4b61-8232-7af9cf1f089c service nova] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.593920] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.845630] env[62385]: DEBUG nova.network.neutron [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: da94a435-b36c-42b2-8c84-564265872048] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 680.935716] env[62385]: DEBUG nova.network.neutron [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: da94a435-b36c-42b2-8c84-564265872048] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.948693] env[62385]: DEBUG oslo_concurrency.lockutils [req-f90f899e-36a6-4251-a0a6-16e59fb7215f req-cb62ef0a-50c3-4b61-8232-7af9cf1f089c service nova] Releasing lock "refresh_cache-7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 680.949967] env[62385]: DEBUG oslo_concurrency.lockutils [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] Acquired lock "refresh_cache-7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.950031] env[62385]: DEBUG nova.network.neutron [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 681.215572] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71c12a34-7ef1-4e19-8ddd-85efe631ce57 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.223729] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2a91c19-f8f0-4a27-98a9-0a273f8e5e75 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.256128] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3cf55aa-ae89-4c84-98ad-1738bcc4a526 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.263710] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45ba279a-6474-4cdc-8063-3781b99e6877 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.276628] env[62385]: DEBUG nova.compute.provider_tree [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 681.442607] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Releasing lock "refresh_cache-da94a435-b36c-42b2-8c84-564265872048" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.442859] env[62385]: DEBUG nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 681.443063] env[62385]: DEBUG nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: da94a435-b36c-42b2-8c84-564265872048] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 681.443240] env[62385]: DEBUG nova.network.neutron [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: da94a435-b36c-42b2-8c84-564265872048] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 681.458780] env[62385]: DEBUG nova.network.neutron [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: da94a435-b36c-42b2-8c84-564265872048] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 681.471258] env[62385]: DEBUG nova.network.neutron [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 681.551029] env[62385]: DEBUG nova.network.neutron [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.779637] env[62385]: DEBUG nova.scheduler.client.report [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 681.859938] env[62385]: DEBUG nova.compute.manager [req-2121a63f-a4f7-4d0f-a211-87140d46ff17 req-a0b63473-9529-4f43-8935-033606b442ea service nova] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Received event network-vif-deleted-0776ce7e-e90a-4326-ac85-6cdc70f2e785 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 681.964843] env[62385]: DEBUG nova.network.neutron [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: da94a435-b36c-42b2-8c84-564265872048] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.053199] env[62385]: DEBUG oslo_concurrency.lockutils [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] Releasing lock "refresh_cache-7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 682.053646] env[62385]: DEBUG nova.compute.manager [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 682.053867] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 682.054160] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-083c064f-6f3c-4cd6-b0b0-318ed6ac9c42 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.063701] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1b943ee-a348-4030-9f31-656258316d86 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.084647] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427 could not be found. [ 682.084886] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 682.085080] env[62385]: INFO nova.compute.manager [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Took 0.03 seconds to destroy the instance on the hypervisor. [ 682.085796] env[62385]: DEBUG oslo.service.loopingcall [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 682.085796] env[62385]: DEBUG nova.compute.manager [-] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 682.085796] env[62385]: DEBUG nova.network.neutron [-] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 682.100673] env[62385]: DEBUG nova.network.neutron [-] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 682.284544] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.978s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.285247] env[62385]: ERROR nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4d8d9700-e4af-4437-88e3-fa5726600f49, please check neutron logs for more information. [ 682.285247] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Traceback (most recent call last): [ 682.285247] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 682.285247] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] self.driver.spawn(context, instance, image_meta, [ 682.285247] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 682.285247] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] self._vmops.spawn(context, instance, image_meta, injected_files, [ 682.285247] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 682.285247] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] vm_ref = self.build_virtual_machine(instance, [ 682.285247] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 682.285247] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] vif_infos = vmwarevif.get_vif_info(self._session, [ 682.285247] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 682.285546] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] for vif in network_info: [ 682.285546] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 682.285546] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] return self._sync_wrapper(fn, *args, **kwargs) [ 682.285546] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 682.285546] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] self.wait() [ 682.285546] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 682.285546] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] self[:] = self._gt.wait() [ 682.285546] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 682.285546] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] return self._exit_event.wait() [ 682.285546] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 682.285546] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] current.throw(*self._exc) [ 682.285546] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 682.285546] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] result = function(*args, **kwargs) [ 682.285868] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 682.285868] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] return func(*args, **kwargs) [ 682.285868] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 682.285868] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] raise e [ 682.285868] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 682.285868] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] nwinfo = self.network_api.allocate_for_instance( [ 682.285868] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 682.285868] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] created_port_ids = self._update_ports_for_instance( [ 682.285868] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 682.285868] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] with excutils.save_and_reraise_exception(): [ 682.285868] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 682.285868] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] self.force_reraise() [ 682.285868] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 682.286201] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] raise self.value [ 682.286201] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 682.286201] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] updated_port = self._update_port( [ 682.286201] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 682.286201] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] _ensure_no_port_binding_failure(port) [ 682.286201] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 682.286201] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] raise exception.PortBindingFailed(port_id=port['id']) [ 682.286201] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] nova.exception.PortBindingFailed: Binding failed for port 4d8d9700-e4af-4437-88e3-fa5726600f49, please check neutron logs for more information. [ 682.286201] env[62385]: ERROR nova.compute.manager [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] [ 682.286201] env[62385]: DEBUG nova.compute.utils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Binding failed for port 4d8d9700-e4af-4437-88e3-fa5726600f49, please check neutron logs for more information. {{(pid=62385) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 682.287662] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.185s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.290789] env[62385]: DEBUG nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Build of instance 7b159d98-9fd2-4eca-9755-cd2cac502798 was re-scheduled: Binding failed for port 4d8d9700-e4af-4437-88e3-fa5726600f49, please check neutron logs for more information. {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 682.290968] env[62385]: DEBUG nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Unplugging VIFs for instance {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 682.291207] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Acquiring lock "refresh_cache-7b159d98-9fd2-4eca-9755-cd2cac502798" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 682.291351] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Acquired lock "refresh_cache-7b159d98-9fd2-4eca-9755-cd2cac502798" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 682.291511] env[62385]: DEBUG nova.network.neutron [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 682.469234] env[62385]: INFO nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: da94a435-b36c-42b2-8c84-564265872048] Took 1.03 seconds to deallocate network for instance. [ 682.603386] env[62385]: DEBUG nova.network.neutron [-] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.814416] env[62385]: DEBUG nova.network.neutron [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 682.928323] env[62385]: DEBUG nova.network.neutron [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.105887] env[62385]: INFO nova.compute.manager [-] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Took 1.02 seconds to deallocate network for instance. [ 683.110047] env[62385]: DEBUG nova.compute.claims [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Aborting claim: {{(pid=62385) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 683.110047] env[62385]: DEBUG oslo_concurrency.lockutils [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 683.221441] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad73774d-eb90-4d37-90ab-cee057d7b005 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.229388] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62ad7f99-c4d0-4fb1-8d3e-b6c786207d58 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.259019] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8d6b5c8-176e-47b3-96f3-97e7dc89dfbc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.266611] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed52c849-b9a8-484e-b2b0-832f9113a5fd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.280763] env[62385]: DEBUG nova.compute.provider_tree [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 683.432369] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Releasing lock "refresh_cache-7b159d98-9fd2-4eca-9755-cd2cac502798" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 683.432692] env[62385]: DEBUG nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 683.432795] env[62385]: DEBUG nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 683.432961] env[62385]: DEBUG nova.network.neutron [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 683.449359] env[62385]: DEBUG nova.network.neutron [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 683.504635] env[62385]: INFO nova.scheduler.client.report [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Deleted allocations for instance da94a435-b36c-42b2-8c84-564265872048 [ 683.788025] env[62385]: DEBUG nova.scheduler.client.report [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 683.952574] env[62385]: DEBUG nova.network.neutron [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.018067] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Lock "da94a435-b36c-42b2-8c84-564265872048" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 122.977s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.290858] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.004s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.291513] env[62385]: ERROR nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6f8b5530-e4f3-4741-a192-0e956321a364, please check neutron logs for more information. [ 684.291513] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Traceback (most recent call last): [ 684.291513] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 684.291513] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] self.driver.spawn(context, instance, image_meta, [ 684.291513] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 684.291513] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 684.291513] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 684.291513] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] vm_ref = self.build_virtual_machine(instance, [ 684.291513] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 684.291513] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] vif_infos = vmwarevif.get_vif_info(self._session, [ 684.291513] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 684.291822] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] for vif in network_info: [ 684.291822] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 684.291822] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] return self._sync_wrapper(fn, *args, **kwargs) [ 684.291822] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 684.291822] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] self.wait() [ 684.291822] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 684.291822] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] self[:] = self._gt.wait() [ 684.291822] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 684.291822] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] return self._exit_event.wait() [ 684.291822] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 684.291822] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] result = hub.switch() [ 684.291822] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 684.291822] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] return self.greenlet.switch() [ 684.292197] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 684.292197] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] result = function(*args, **kwargs) [ 684.292197] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 684.292197] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] return func(*args, **kwargs) [ 684.292197] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 684.292197] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] raise e [ 684.292197] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 684.292197] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] nwinfo = self.network_api.allocate_for_instance( [ 684.292197] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 684.292197] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] created_port_ids = self._update_ports_for_instance( [ 684.292197] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 684.292197] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] with excutils.save_and_reraise_exception(): [ 684.292197] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.292514] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] self.force_reraise() [ 684.292514] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.292514] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] raise self.value [ 684.292514] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 684.292514] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] updated_port = self._update_port( [ 684.292514] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.292514] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] _ensure_no_port_binding_failure(port) [ 684.292514] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.292514] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] raise exception.PortBindingFailed(port_id=port['id']) [ 684.292514] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] nova.exception.PortBindingFailed: Binding failed for port 6f8b5530-e4f3-4741-a192-0e956321a364, please check neutron logs for more information. [ 684.292514] env[62385]: ERROR nova.compute.manager [instance: 608c13da-6625-417d-a7b3-10821ae638d8] [ 684.292782] env[62385]: DEBUG nova.compute.utils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Binding failed for port 6f8b5530-e4f3-4741-a192-0e956321a364, please check neutron logs for more information. {{(pid=62385) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 684.293894] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.780s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 684.295384] env[62385]: INFO nova.compute.claims [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 684.303354] env[62385]: DEBUG nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Build of instance 608c13da-6625-417d-a7b3-10821ae638d8 was re-scheduled: Binding failed for port 6f8b5530-e4f3-4741-a192-0e956321a364, please check neutron logs for more information. {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 684.303828] env[62385]: DEBUG nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Unplugging VIFs for instance {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 684.304120] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Acquiring lock "refresh_cache-608c13da-6625-417d-a7b3-10821ae638d8" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.304213] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Acquired lock "refresh_cache-608c13da-6625-417d-a7b3-10821ae638d8" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.304375] env[62385]: DEBUG nova.network.neutron [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 684.458401] env[62385]: INFO nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 7b159d98-9fd2-4eca-9755-cd2cac502798] Took 1.03 seconds to deallocate network for instance. [ 684.523714] env[62385]: DEBUG nova.compute.manager [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 684.829221] env[62385]: DEBUG nova.network.neutron [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 684.950123] env[62385]: DEBUG nova.network.neutron [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.045976] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 685.453031] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Releasing lock "refresh_cache-608c13da-6625-417d-a7b3-10821ae638d8" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.453270] env[62385]: DEBUG nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 685.453455] env[62385]: DEBUG nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 685.453644] env[62385]: DEBUG nova.network.neutron [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 685.477422] env[62385]: DEBUG nova.network.neutron [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 685.488708] env[62385]: INFO nova.scheduler.client.report [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Deleted allocations for instance 7b159d98-9fd2-4eca-9755-cd2cac502798 [ 685.707590] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1ba8506-dc71-4be9-a616-1a37a322ad27 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.716031] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a79d85b-839b-445c-b3c0-6f56bebb7ccc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.746247] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73206c8b-99f6-4a69-90e0-178e8a9e2c51 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.753557] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b08d197-a5b5-4fad-90ea-ef8297fe286d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.766433] env[62385]: DEBUG nova.compute.provider_tree [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 685.802867] env[62385]: DEBUG oslo_concurrency.lockutils [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "0d823fcc-cbba-41f6-a47c-2f57d3e79948" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 685.803106] env[62385]: DEBUG oslo_concurrency.lockutils [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "0d823fcc-cbba-41f6-a47c-2f57d3e79948" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 685.983946] env[62385]: DEBUG nova.network.neutron [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.000116] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Lock "7b159d98-9fd2-4eca-9755-cd2cac502798" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 124.930s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.269694] env[62385]: DEBUG nova.scheduler.client.report [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 686.487086] env[62385]: INFO nova.compute.manager [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] [instance: 608c13da-6625-417d-a7b3-10821ae638d8] Took 1.03 seconds to deallocate network for instance. [ 686.502521] env[62385]: DEBUG nova.compute.manager [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 686.777781] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.484s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.778109] env[62385]: DEBUG nova.compute.manager [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 686.780565] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.380s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.783150] env[62385]: INFO nova.compute.claims [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 687.024803] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.286889] env[62385]: DEBUG nova.compute.utils [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 687.290943] env[62385]: DEBUG nova.compute.manager [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 687.290943] env[62385]: DEBUG nova.network.neutron [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 687.355234] env[62385]: DEBUG nova.policy [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'de410965670c4db9a14c18fb998984cd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0e705544b9134581b8611bf53c948578', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 687.520053] env[62385]: INFO nova.scheduler.client.report [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Deleted allocations for instance 608c13da-6625-417d-a7b3-10821ae638d8 [ 687.748616] env[62385]: DEBUG nova.network.neutron [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Successfully created port: 20df7395-1da3-486e-a35e-0279141f722a {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 687.791345] env[62385]: DEBUG nova.compute.manager [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 688.028313] env[62385]: DEBUG oslo_concurrency.lockutils [None req-70927004-fe9c-4ebc-b2ec-515a6ddda5eb tempest-ListServersNegativeTestJSON-2133975674 tempest-ListServersNegativeTestJSON-2133975674-project-member] Lock "608c13da-6625-417d-a7b3-10821ae638d8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 126.921s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 688.131604] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05b86c6d-a812-461d-ac47-98a20017c945 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.139489] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18b5d17b-19d3-4f25-924b-9ae7754bd0af {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.169694] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d075a61-1fde-4118-9b55-611a5e8863a6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.177268] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d9b5e1c-4bc5-4570-9cc6-be33aaeff605 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.191629] env[62385]: DEBUG nova.compute.provider_tree [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 688.530897] env[62385]: DEBUG nova.compute.manager [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 688.694420] env[62385]: DEBUG nova.scheduler.client.report [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 688.800945] env[62385]: DEBUG nova.compute.manager [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 688.828930] env[62385]: DEBUG nova.virt.hardware [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 688.829105] env[62385]: DEBUG nova.virt.hardware [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 688.829267] env[62385]: DEBUG nova.virt.hardware [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 688.829449] env[62385]: DEBUG nova.virt.hardware [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 688.829593] env[62385]: DEBUG nova.virt.hardware [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 688.829735] env[62385]: DEBUG nova.virt.hardware [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 688.829941] env[62385]: DEBUG nova.virt.hardware [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 688.830193] env[62385]: DEBUG nova.virt.hardware [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 688.830378] env[62385]: DEBUG nova.virt.hardware [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 688.830542] env[62385]: DEBUG nova.virt.hardware [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 688.830710] env[62385]: DEBUG nova.virt.hardware [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 688.831592] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d54cb65-b487-485e-bed2-0fa83ace6dcf {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.840157] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44971aa2-a08b-4efc-9560-0ff352810ec1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.056621] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 689.085381] env[62385]: DEBUG nova.compute.manager [req-21f3057e-4899-460d-bd91-433d6ea88fcf req-037c75c9-6ce9-4eaa-97f5-52e6bab445d5 service nova] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Received event network-changed-20df7395-1da3-486e-a35e-0279141f722a {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 689.085381] env[62385]: DEBUG nova.compute.manager [req-21f3057e-4899-460d-bd91-433d6ea88fcf req-037c75c9-6ce9-4eaa-97f5-52e6bab445d5 service nova] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Refreshing instance network info cache due to event network-changed-20df7395-1da3-486e-a35e-0279141f722a. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 689.085381] env[62385]: DEBUG oslo_concurrency.lockutils [req-21f3057e-4899-460d-bd91-433d6ea88fcf req-037c75c9-6ce9-4eaa-97f5-52e6bab445d5 service nova] Acquiring lock "refresh_cache-1b459bac-e27d-4af5-bfc2-49898f5ced25" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 689.085666] env[62385]: DEBUG oslo_concurrency.lockutils [req-21f3057e-4899-460d-bd91-433d6ea88fcf req-037c75c9-6ce9-4eaa-97f5-52e6bab445d5 service nova] Acquired lock "refresh_cache-1b459bac-e27d-4af5-bfc2-49898f5ced25" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.085698] env[62385]: DEBUG nova.network.neutron [req-21f3057e-4899-460d-bd91-433d6ea88fcf req-037c75c9-6ce9-4eaa-97f5-52e6bab445d5 service nova] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Refreshing network info cache for port 20df7395-1da3-486e-a35e-0279141f722a {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 689.202134] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.421s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 689.202668] env[62385]: DEBUG nova.compute.manager [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 689.209055] env[62385]: DEBUG oslo_concurrency.lockutils [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.144s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 689.463108] env[62385]: ERROR nova.compute.manager [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 20df7395-1da3-486e-a35e-0279141f722a, please check neutron logs for more information. [ 689.463108] env[62385]: ERROR nova.compute.manager Traceback (most recent call last): [ 689.463108] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 689.463108] env[62385]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 689.463108] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 689.463108] env[62385]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 689.463108] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 689.463108] env[62385]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 689.463108] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.463108] env[62385]: ERROR nova.compute.manager self.force_reraise() [ 689.463108] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.463108] env[62385]: ERROR nova.compute.manager raise self.value [ 689.463108] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 689.463108] env[62385]: ERROR nova.compute.manager updated_port = self._update_port( [ 689.463108] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.463108] env[62385]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 689.463607] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.463607] env[62385]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 689.463607] env[62385]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 20df7395-1da3-486e-a35e-0279141f722a, please check neutron logs for more information. [ 689.463607] env[62385]: ERROR nova.compute.manager [ 689.463607] env[62385]: Traceback (most recent call last): [ 689.463607] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 689.463607] env[62385]: listener.cb(fileno) [ 689.463607] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 689.463607] env[62385]: result = function(*args, **kwargs) [ 689.463607] env[62385]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 689.463607] env[62385]: return func(*args, **kwargs) [ 689.463607] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 689.463607] env[62385]: raise e [ 689.463607] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 689.463607] env[62385]: nwinfo = self.network_api.allocate_for_instance( [ 689.463607] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 689.463607] env[62385]: created_port_ids = self._update_ports_for_instance( [ 689.463607] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 689.463607] env[62385]: with excutils.save_and_reraise_exception(): [ 689.463607] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.463607] env[62385]: self.force_reraise() [ 689.463607] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.463607] env[62385]: raise self.value [ 689.463607] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 689.463607] env[62385]: updated_port = self._update_port( [ 689.463607] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.463607] env[62385]: _ensure_no_port_binding_failure(port) [ 689.463607] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.463607] env[62385]: raise exception.PortBindingFailed(port_id=port['id']) [ 689.464338] env[62385]: nova.exception.PortBindingFailed: Binding failed for port 20df7395-1da3-486e-a35e-0279141f722a, please check neutron logs for more information. [ 689.464338] env[62385]: Removing descriptor: 19 [ 689.464338] env[62385]: ERROR nova.compute.manager [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 20df7395-1da3-486e-a35e-0279141f722a, please check neutron logs for more information. [ 689.464338] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Traceback (most recent call last): [ 689.464338] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 689.464338] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] yield resources [ 689.464338] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 689.464338] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] self.driver.spawn(context, instance, image_meta, [ 689.464338] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 689.464338] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] self._vmops.spawn(context, instance, image_meta, injected_files, [ 689.464338] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 689.464338] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] vm_ref = self.build_virtual_machine(instance, [ 689.464653] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 689.464653] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] vif_infos = vmwarevif.get_vif_info(self._session, [ 689.464653] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 689.464653] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] for vif in network_info: [ 689.464653] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 689.464653] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] return self._sync_wrapper(fn, *args, **kwargs) [ 689.464653] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 689.464653] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] self.wait() [ 689.464653] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 689.464653] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] self[:] = self._gt.wait() [ 689.464653] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 689.464653] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] return self._exit_event.wait() [ 689.464653] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 689.464996] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] result = hub.switch() [ 689.464996] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 689.464996] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] return self.greenlet.switch() [ 689.464996] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 689.464996] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] result = function(*args, **kwargs) [ 689.464996] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 689.464996] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] return func(*args, **kwargs) [ 689.464996] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 689.464996] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] raise e [ 689.464996] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 689.464996] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] nwinfo = self.network_api.allocate_for_instance( [ 689.464996] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 689.464996] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] created_port_ids = self._update_ports_for_instance( [ 689.465355] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 689.465355] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] with excutils.save_and_reraise_exception(): [ 689.465355] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.465355] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] self.force_reraise() [ 689.465355] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.465355] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] raise self.value [ 689.465355] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 689.465355] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] updated_port = self._update_port( [ 689.465355] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.465355] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] _ensure_no_port_binding_failure(port) [ 689.465355] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.465355] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] raise exception.PortBindingFailed(port_id=port['id']) [ 689.465671] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] nova.exception.PortBindingFailed: Binding failed for port 20df7395-1da3-486e-a35e-0279141f722a, please check neutron logs for more information. [ 689.465671] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] [ 689.465671] env[62385]: INFO nova.compute.manager [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Terminating instance [ 689.466341] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] Acquiring lock "refresh_cache-1b459bac-e27d-4af5-bfc2-49898f5ced25" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 689.605907] env[62385]: DEBUG nova.network.neutron [req-21f3057e-4899-460d-bd91-433d6ea88fcf req-037c75c9-6ce9-4eaa-97f5-52e6bab445d5 service nova] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 689.710331] env[62385]: DEBUG nova.compute.utils [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 689.714684] env[62385]: DEBUG nova.compute.manager [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 689.714850] env[62385]: DEBUG nova.network.neutron [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 689.748178] env[62385]: DEBUG nova.network.neutron [req-21f3057e-4899-460d-bd91-433d6ea88fcf req-037c75c9-6ce9-4eaa-97f5-52e6bab445d5 service nova] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.784629] env[62385]: DEBUG nova.policy [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9d0da5e5b9f54856ad914b313719b824', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '26ce94e5e21c4a64aebba9fb08918678', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 690.123371] env[62385]: DEBUG nova.network.neutron [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Successfully created port: 6091e7b7-501c-4419-b688-843834c15097 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 690.130782] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-660585c8-14fd-40d4-9a2a-a2bdeb3a1e7e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.139390] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-232b9dc7-f3a0-437c-9d23-3d98399d13d8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.173871] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91fd75bc-95b6-4971-a9c3-06133dd990f9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.182183] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c703e0c2-4bba-412f-bb2d-23f83db910fe {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.196752] env[62385]: DEBUG nova.compute.provider_tree [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 690.216122] env[62385]: DEBUG nova.compute.manager [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 690.254183] env[62385]: DEBUG oslo_concurrency.lockutils [req-21f3057e-4899-460d-bd91-433d6ea88fcf req-037c75c9-6ce9-4eaa-97f5-52e6bab445d5 service nova] Releasing lock "refresh_cache-1b459bac-e27d-4af5-bfc2-49898f5ced25" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.254614] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] Acquired lock "refresh_cache-1b459bac-e27d-4af5-bfc2-49898f5ced25" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.254842] env[62385]: DEBUG nova.network.neutron [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 690.699841] env[62385]: DEBUG nova.scheduler.client.report [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 690.778537] env[62385]: DEBUG nova.network.neutron [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 690.806723] env[62385]: DEBUG nova.compute.manager [req-d5306855-015f-4dd1-b83a-444b6881c550 req-bfd20c31-d678-4d68-8d39-8a9e65d34465 service nova] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Received event network-changed-6091e7b7-501c-4419-b688-843834c15097 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 690.806915] env[62385]: DEBUG nova.compute.manager [req-d5306855-015f-4dd1-b83a-444b6881c550 req-bfd20c31-d678-4d68-8d39-8a9e65d34465 service nova] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Refreshing instance network info cache due to event network-changed-6091e7b7-501c-4419-b688-843834c15097. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 690.807142] env[62385]: DEBUG oslo_concurrency.lockutils [req-d5306855-015f-4dd1-b83a-444b6881c550 req-bfd20c31-d678-4d68-8d39-8a9e65d34465 service nova] Acquiring lock "refresh_cache-d3224596-4e38-4c9c-abd9-59a34bbf59c7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 690.807314] env[62385]: DEBUG oslo_concurrency.lockutils [req-d5306855-015f-4dd1-b83a-444b6881c550 req-bfd20c31-d678-4d68-8d39-8a9e65d34465 service nova] Acquired lock "refresh_cache-d3224596-4e38-4c9c-abd9-59a34bbf59c7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.807481] env[62385]: DEBUG nova.network.neutron [req-d5306855-015f-4dd1-b83a-444b6881c550 req-bfd20c31-d678-4d68-8d39-8a9e65d34465 service nova] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Refreshing network info cache for port 6091e7b7-501c-4419-b688-843834c15097 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 690.912501] env[62385]: DEBUG nova.network.neutron [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.003270] env[62385]: ERROR nova.compute.manager [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6091e7b7-501c-4419-b688-843834c15097, please check neutron logs for more information. [ 691.003270] env[62385]: ERROR nova.compute.manager Traceback (most recent call last): [ 691.003270] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 691.003270] env[62385]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 691.003270] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 691.003270] env[62385]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 691.003270] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 691.003270] env[62385]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 691.003270] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 691.003270] env[62385]: ERROR nova.compute.manager self.force_reraise() [ 691.003270] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 691.003270] env[62385]: ERROR nova.compute.manager raise self.value [ 691.003270] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 691.003270] env[62385]: ERROR nova.compute.manager updated_port = self._update_port( [ 691.003270] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 691.003270] env[62385]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 691.003703] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 691.003703] env[62385]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 691.003703] env[62385]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6091e7b7-501c-4419-b688-843834c15097, please check neutron logs for more information. [ 691.003703] env[62385]: ERROR nova.compute.manager [ 691.003703] env[62385]: Traceback (most recent call last): [ 691.003703] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 691.003703] env[62385]: listener.cb(fileno) [ 691.003703] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 691.003703] env[62385]: result = function(*args, **kwargs) [ 691.003703] env[62385]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 691.003703] env[62385]: return func(*args, **kwargs) [ 691.003703] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 691.003703] env[62385]: raise e [ 691.003703] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 691.003703] env[62385]: nwinfo = self.network_api.allocate_for_instance( [ 691.003703] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 691.003703] env[62385]: created_port_ids = self._update_ports_for_instance( [ 691.003703] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 691.003703] env[62385]: with excutils.save_and_reraise_exception(): [ 691.003703] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 691.003703] env[62385]: self.force_reraise() [ 691.003703] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 691.003703] env[62385]: raise self.value [ 691.003703] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 691.003703] env[62385]: updated_port = self._update_port( [ 691.003703] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 691.003703] env[62385]: _ensure_no_port_binding_failure(port) [ 691.003703] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 691.003703] env[62385]: raise exception.PortBindingFailed(port_id=port['id']) [ 691.004409] env[62385]: nova.exception.PortBindingFailed: Binding failed for port 6091e7b7-501c-4419-b688-843834c15097, please check neutron logs for more information. [ 691.004409] env[62385]: Removing descriptor: 17 [ 691.205698] env[62385]: DEBUG oslo_concurrency.lockutils [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 691.206372] env[62385]: ERROR nova.compute.manager [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ca62abe6-ab3c-4b9f-ab5e-ad10b81e8be9, please check neutron logs for more information. [ 691.206372] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Traceback (most recent call last): [ 691.206372] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 691.206372] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] self.driver.spawn(context, instance, image_meta, [ 691.206372] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 691.206372] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 691.206372] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 691.206372] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] vm_ref = self.build_virtual_machine(instance, [ 691.206372] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 691.206372] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] vif_infos = vmwarevif.get_vif_info(self._session, [ 691.206372] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 691.206701] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] for vif in network_info: [ 691.206701] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 691.206701] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] return self._sync_wrapper(fn, *args, **kwargs) [ 691.206701] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 691.206701] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] self.wait() [ 691.206701] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 691.206701] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] self[:] = self._gt.wait() [ 691.206701] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 691.206701] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] return self._exit_event.wait() [ 691.206701] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 691.206701] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] result = hub.switch() [ 691.206701] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 691.206701] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] return self.greenlet.switch() [ 691.207029] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 691.207029] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] result = function(*args, **kwargs) [ 691.207029] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 691.207029] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] return func(*args, **kwargs) [ 691.207029] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 691.207029] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] raise e [ 691.207029] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 691.207029] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] nwinfo = self.network_api.allocate_for_instance( [ 691.207029] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 691.207029] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] created_port_ids = self._update_ports_for_instance( [ 691.207029] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 691.207029] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] with excutils.save_and_reraise_exception(): [ 691.207029] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 691.207394] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] self.force_reraise() [ 691.207394] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 691.207394] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] raise self.value [ 691.207394] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 691.207394] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] updated_port = self._update_port( [ 691.207394] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 691.207394] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] _ensure_no_port_binding_failure(port) [ 691.207394] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 691.207394] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] raise exception.PortBindingFailed(port_id=port['id']) [ 691.207394] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] nova.exception.PortBindingFailed: Binding failed for port ca62abe6-ab3c-4b9f-ab5e-ad10b81e8be9, please check neutron logs for more information. [ 691.207394] env[62385]: ERROR nova.compute.manager [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] [ 691.207674] env[62385]: DEBUG nova.compute.utils [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Binding failed for port ca62abe6-ab3c-4b9f-ab5e-ad10b81e8be9, please check neutron logs for more information. {{(pid=62385) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 691.208610] env[62385]: DEBUG oslo_concurrency.lockutils [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.129s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 691.210243] env[62385]: INFO nova.compute.claims [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 691.213167] env[62385]: DEBUG nova.compute.manager [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Build of instance 6e0dd86b-d557-4e7a-8520-9ffbd291128b was re-scheduled: Binding failed for port ca62abe6-ab3c-4b9f-ab5e-ad10b81e8be9, please check neutron logs for more information. {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 691.213651] env[62385]: DEBUG nova.compute.manager [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Unplugging VIFs for instance {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 691.214178] env[62385]: DEBUG oslo_concurrency.lockutils [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Acquiring lock "refresh_cache-6e0dd86b-d557-4e7a-8520-9ffbd291128b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 691.214386] env[62385]: DEBUG oslo_concurrency.lockutils [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Acquired lock "refresh_cache-6e0dd86b-d557-4e7a-8520-9ffbd291128b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 691.214590] env[62385]: DEBUG nova.network.neutron [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 691.225565] env[62385]: DEBUG nova.compute.manager [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 691.256693] env[62385]: DEBUG nova.virt.hardware [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:18:30Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='d3d3263a-e792-4189-a0dd-67b3c0f08e38',id=35,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1552947710',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 691.256934] env[62385]: DEBUG nova.virt.hardware [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 691.257105] env[62385]: DEBUG nova.virt.hardware [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 691.257302] env[62385]: DEBUG nova.virt.hardware [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 691.257476] env[62385]: DEBUG nova.virt.hardware [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 691.258184] env[62385]: DEBUG nova.virt.hardware [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 691.258509] env[62385]: DEBUG nova.virt.hardware [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 691.258681] env[62385]: DEBUG nova.virt.hardware [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 691.258854] env[62385]: DEBUG nova.virt.hardware [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 691.259029] env[62385]: DEBUG nova.virt.hardware [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 691.259406] env[62385]: DEBUG nova.virt.hardware [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 691.260109] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34a39276-058f-49b2-9c76-a0bfd80fde93 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.270241] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17646e25-a258-462c-9fce-b97d925192fc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.284769] env[62385]: ERROR nova.compute.manager [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6091e7b7-501c-4419-b688-843834c15097, please check neutron logs for more information. [ 691.284769] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Traceback (most recent call last): [ 691.284769] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 691.284769] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] yield resources [ 691.284769] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 691.284769] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] self.driver.spawn(context, instance, image_meta, [ 691.284769] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 691.284769] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 691.284769] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 691.284769] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] vm_ref = self.build_virtual_machine(instance, [ 691.284769] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 691.285223] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] vif_infos = vmwarevif.get_vif_info(self._session, [ 691.285223] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 691.285223] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] for vif in network_info: [ 691.285223] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 691.285223] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] return self._sync_wrapper(fn, *args, **kwargs) [ 691.285223] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 691.285223] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] self.wait() [ 691.285223] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 691.285223] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] self[:] = self._gt.wait() [ 691.285223] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 691.285223] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] return self._exit_event.wait() [ 691.285223] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 691.285223] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] current.throw(*self._exc) [ 691.285550] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 691.285550] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] result = function(*args, **kwargs) [ 691.285550] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 691.285550] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] return func(*args, **kwargs) [ 691.285550] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 691.285550] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] raise e [ 691.285550] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 691.285550] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] nwinfo = self.network_api.allocate_for_instance( [ 691.285550] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 691.285550] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] created_port_ids = self._update_ports_for_instance( [ 691.285550] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 691.285550] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] with excutils.save_and_reraise_exception(): [ 691.285550] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 691.285923] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] self.force_reraise() [ 691.285923] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 691.285923] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] raise self.value [ 691.285923] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 691.285923] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] updated_port = self._update_port( [ 691.285923] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 691.285923] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] _ensure_no_port_binding_failure(port) [ 691.285923] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 691.285923] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] raise exception.PortBindingFailed(port_id=port['id']) [ 691.285923] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] nova.exception.PortBindingFailed: Binding failed for port 6091e7b7-501c-4419-b688-843834c15097, please check neutron logs for more information. [ 691.285923] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] [ 691.285923] env[62385]: INFO nova.compute.manager [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Terminating instance [ 691.287180] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Acquiring lock "refresh_cache-d3224596-4e38-4c9c-abd9-59a34bbf59c7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 691.309385] env[62385]: DEBUG nova.compute.manager [req-abeb7660-5366-4611-aa33-6c40bef656a4 req-82913ae2-39e0-4ce1-84a7-a791935a8207 service nova] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Received event network-vif-deleted-20df7395-1da3-486e-a35e-0279141f722a {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 691.330485] env[62385]: DEBUG nova.network.neutron [req-d5306855-015f-4dd1-b83a-444b6881c550 req-bfd20c31-d678-4d68-8d39-8a9e65d34465 service nova] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 691.393117] env[62385]: DEBUG nova.network.neutron [req-d5306855-015f-4dd1-b83a-444b6881c550 req-bfd20c31-d678-4d68-8d39-8a9e65d34465 service nova] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.413860] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] Releasing lock "refresh_cache-1b459bac-e27d-4af5-bfc2-49898f5ced25" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.414739] env[62385]: DEBUG nova.compute.manager [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 691.415138] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 691.415368] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-27fac7e7-246e-4b65-a9ef-afe2a7b9fcdc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.425185] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49a70a8c-e197-43ff-a7bb-c70d96495422 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.446863] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1b459bac-e27d-4af5-bfc2-49898f5ced25 could not be found. [ 691.447049] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 691.447235] env[62385]: INFO nova.compute.manager [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Took 0.03 seconds to destroy the instance on the hypervisor. [ 691.447539] env[62385]: DEBUG oslo.service.loopingcall [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 691.447757] env[62385]: DEBUG nova.compute.manager [-] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 691.447853] env[62385]: DEBUG nova.network.neutron [-] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 691.462786] env[62385]: DEBUG nova.network.neutron [-] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 691.736009] env[62385]: DEBUG nova.network.neutron [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 691.804279] env[62385]: DEBUG nova.network.neutron [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.894744] env[62385]: DEBUG oslo_concurrency.lockutils [req-d5306855-015f-4dd1-b83a-444b6881c550 req-bfd20c31-d678-4d68-8d39-8a9e65d34465 service nova] Releasing lock "refresh_cache-d3224596-4e38-4c9c-abd9-59a34bbf59c7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.896067] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Acquired lock "refresh_cache-d3224596-4e38-4c9c-abd9-59a34bbf59c7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 691.896241] env[62385]: DEBUG nova.network.neutron [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 691.968875] env[62385]: DEBUG nova.network.neutron [-] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.307929] env[62385]: DEBUG oslo_concurrency.lockutils [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Releasing lock "refresh_cache-6e0dd86b-d557-4e7a-8520-9ffbd291128b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 692.307929] env[62385]: DEBUG nova.compute.manager [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 692.307929] env[62385]: DEBUG nova.compute.manager [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 692.307929] env[62385]: DEBUG nova.network.neutron [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 692.323486] env[62385]: DEBUG nova.network.neutron [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 692.421132] env[62385]: DEBUG nova.network.neutron [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 692.471049] env[62385]: INFO nova.compute.manager [-] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Took 1.02 seconds to deallocate network for instance. [ 692.476988] env[62385]: DEBUG nova.compute.claims [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Aborting claim: {{(pid=62385) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 692.477185] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.487501] env[62385]: DEBUG nova.network.neutron [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.617647] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ae8dcd9-39fd-4d76-9ccd-b06fda3b4630 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.625870] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82bd37b0-e2b7-4586-ab7e-3470911c22f7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.655523] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6042793e-349b-48e6-a4ec-867367420c74 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.663258] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-120a1c22-4a25-4c24-9e66-44a91a7697a4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.676348] env[62385]: DEBUG nova.compute.provider_tree [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 692.828911] env[62385]: DEBUG nova.network.neutron [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.836325] env[62385]: DEBUG nova.compute.manager [req-68e9e202-e220-4b67-b47d-ddbe56ed754d req-dc7428df-cb6c-4dbb-9871-5ae604ba1ff1 service nova] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Received event network-vif-deleted-6091e7b7-501c-4419-b688-843834c15097 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 692.994018] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Releasing lock "refresh_cache-d3224596-4e38-4c9c-abd9-59a34bbf59c7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 692.994018] env[62385]: DEBUG nova.compute.manager [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 692.994018] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 692.994018] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-847354d2-723e-43bd-8350-03ef0eb0deea {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.005782] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39ec3883-dc02-4cfc-a8c4-e47c79896cc0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.032239] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d3224596-4e38-4c9c-abd9-59a34bbf59c7 could not be found. [ 693.032468] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 693.032649] env[62385]: INFO nova.compute.manager [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Took 0.04 seconds to destroy the instance on the hypervisor. [ 693.032894] env[62385]: DEBUG oslo.service.loopingcall [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 693.033221] env[62385]: DEBUG nova.compute.manager [-] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 693.033221] env[62385]: DEBUG nova.network.neutron [-] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 693.048018] env[62385]: DEBUG nova.network.neutron [-] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 693.185034] env[62385]: DEBUG nova.scheduler.client.report [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 693.333949] env[62385]: INFO nova.compute.manager [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: 6e0dd86b-d557-4e7a-8520-9ffbd291128b] Took 1.02 seconds to deallocate network for instance. [ 693.551230] env[62385]: DEBUG nova.network.neutron [-] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 693.690199] env[62385]: DEBUG oslo_concurrency.lockutils [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.479s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 693.690199] env[62385]: DEBUG nova.compute.manager [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 693.692418] env[62385]: DEBUG oslo_concurrency.lockutils [None req-50e22468-1913-42eb-9f13-0da6a2f22848 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.932s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 693.693201] env[62385]: DEBUG nova.objects.instance [None req-50e22468-1913-42eb-9f13-0da6a2f22848 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Lazy-loading 'resources' on Instance uuid 0ba32627-c646-477d-9062-b84624d01e48 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 694.054390] env[62385]: INFO nova.compute.manager [-] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Took 1.02 seconds to deallocate network for instance. [ 694.056952] env[62385]: DEBUG nova.compute.claims [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Aborting claim: {{(pid=62385) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 694.057173] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 694.197378] env[62385]: DEBUG nova.compute.utils [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 694.202162] env[62385]: DEBUG nova.compute.manager [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 694.202162] env[62385]: DEBUG nova.network.neutron [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 694.261609] env[62385]: DEBUG nova.policy [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '73a2a8bde6624453be40a13c97d4f5cb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'edf2b9d9c6914a59bf8fe00247f35c4b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 694.364590] env[62385]: INFO nova.scheduler.client.report [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Deleted allocations for instance 6e0dd86b-d557-4e7a-8520-9ffbd291128b [ 694.594608] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9899699-2e66-48c9-aa02-272d711d6d3b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.602211] env[62385]: DEBUG nova.network.neutron [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Successfully created port: ad28927c-3e0a-47c2-b304-43949d2afc0d {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 694.605367] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98752362-45d9-4f17-ae72-b51f22dcbf28 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.638472] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfc4a528-1564-4082-918f-26e6603e59fd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.646803] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2029d29d-e556-4c41-83cb-ba86b8de6249 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.659839] env[62385]: DEBUG nova.compute.provider_tree [None req-50e22468-1913-42eb-9f13-0da6a2f22848 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 694.704541] env[62385]: DEBUG nova.compute.manager [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 694.833942] env[62385]: DEBUG oslo_concurrency.lockutils [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Acquiring lock "b4970e1a-2b29-44e9-b79e-cda3c32dddcd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 694.833942] env[62385]: DEBUG oslo_concurrency.lockutils [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Lock "b4970e1a-2b29-44e9-b79e-cda3c32dddcd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 694.877653] env[62385]: DEBUG oslo_concurrency.lockutils [None req-28dab765-8817-4106-9769-ff03feff0fed tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Lock "6e0dd86b-d557-4e7a-8520-9ffbd291128b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 132.833s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.163321] env[62385]: DEBUG nova.scheduler.client.report [None req-50e22468-1913-42eb-9f13-0da6a2f22848 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 695.371100] env[62385]: DEBUG nova.compute.manager [req-a15187f8-1588-4024-bf64-d4532f498a15 req-df2c2d2c-83cb-4598-985f-2aa6c662b126 service nova] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Received event network-changed-ad28927c-3e0a-47c2-b304-43949d2afc0d {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 695.371294] env[62385]: DEBUG nova.compute.manager [req-a15187f8-1588-4024-bf64-d4532f498a15 req-df2c2d2c-83cb-4598-985f-2aa6c662b126 service nova] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Refreshing instance network info cache due to event network-changed-ad28927c-3e0a-47c2-b304-43949d2afc0d. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 695.371508] env[62385]: DEBUG oslo_concurrency.lockutils [req-a15187f8-1588-4024-bf64-d4532f498a15 req-df2c2d2c-83cb-4598-985f-2aa6c662b126 service nova] Acquiring lock "refresh_cache-4c96b9c6-40b9-4dea-8dd3-5e8e6372c261" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 695.371717] env[62385]: DEBUG oslo_concurrency.lockutils [req-a15187f8-1588-4024-bf64-d4532f498a15 req-df2c2d2c-83cb-4598-985f-2aa6c662b126 service nova] Acquired lock "refresh_cache-4c96b9c6-40b9-4dea-8dd3-5e8e6372c261" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.371826] env[62385]: DEBUG nova.network.neutron [req-a15187f8-1588-4024-bf64-d4532f498a15 req-df2c2d2c-83cb-4598-985f-2aa6c662b126 service nova] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Refreshing network info cache for port ad28927c-3e0a-47c2-b304-43949d2afc0d {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 695.380533] env[62385]: DEBUG nova.compute.manager [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 695.525054] env[62385]: ERROR nova.compute.manager [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ad28927c-3e0a-47c2-b304-43949d2afc0d, please check neutron logs for more information. [ 695.525054] env[62385]: ERROR nova.compute.manager Traceback (most recent call last): [ 695.525054] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 695.525054] env[62385]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 695.525054] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 695.525054] env[62385]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 695.525054] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 695.525054] env[62385]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 695.525054] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 695.525054] env[62385]: ERROR nova.compute.manager self.force_reraise() [ 695.525054] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 695.525054] env[62385]: ERROR nova.compute.manager raise self.value [ 695.525054] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 695.525054] env[62385]: ERROR nova.compute.manager updated_port = self._update_port( [ 695.525054] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 695.525054] env[62385]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 695.525520] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 695.525520] env[62385]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 695.525520] env[62385]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ad28927c-3e0a-47c2-b304-43949d2afc0d, please check neutron logs for more information. [ 695.525520] env[62385]: ERROR nova.compute.manager [ 695.525520] env[62385]: Traceback (most recent call last): [ 695.525520] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 695.525520] env[62385]: listener.cb(fileno) [ 695.525520] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 695.525520] env[62385]: result = function(*args, **kwargs) [ 695.525520] env[62385]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 695.525520] env[62385]: return func(*args, **kwargs) [ 695.525520] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 695.525520] env[62385]: raise e [ 695.525520] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 695.525520] env[62385]: nwinfo = self.network_api.allocate_for_instance( [ 695.525520] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 695.525520] env[62385]: created_port_ids = self._update_ports_for_instance( [ 695.525520] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 695.525520] env[62385]: with excutils.save_and_reraise_exception(): [ 695.525520] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 695.525520] env[62385]: self.force_reraise() [ 695.525520] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 695.525520] env[62385]: raise self.value [ 695.525520] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 695.525520] env[62385]: updated_port = self._update_port( [ 695.525520] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 695.525520] env[62385]: _ensure_no_port_binding_failure(port) [ 695.525520] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 695.525520] env[62385]: raise exception.PortBindingFailed(port_id=port['id']) [ 695.526393] env[62385]: nova.exception.PortBindingFailed: Binding failed for port ad28927c-3e0a-47c2-b304-43949d2afc0d, please check neutron logs for more information. [ 695.526393] env[62385]: Removing descriptor: 17 [ 695.669745] env[62385]: DEBUG oslo_concurrency.lockutils [None req-50e22468-1913-42eb-9f13-0da6a2f22848 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.977s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.671998] env[62385]: DEBUG oslo_concurrency.lockutils [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.981s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 695.694508] env[62385]: INFO nova.scheduler.client.report [None req-50e22468-1913-42eb-9f13-0da6a2f22848 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Deleted allocations for instance 0ba32627-c646-477d-9062-b84624d01e48 [ 695.722154] env[62385]: DEBUG nova.compute.manager [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 695.743011] env[62385]: DEBUG nova.virt.hardware [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 695.743269] env[62385]: DEBUG nova.virt.hardware [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 695.743423] env[62385]: DEBUG nova.virt.hardware [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 695.743604] env[62385]: DEBUG nova.virt.hardware [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 695.743746] env[62385]: DEBUG nova.virt.hardware [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 695.743891] env[62385]: DEBUG nova.virt.hardware [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 695.744336] env[62385]: DEBUG nova.virt.hardware [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 695.744336] env[62385]: DEBUG nova.virt.hardware [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 695.744467] env[62385]: DEBUG nova.virt.hardware [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 695.744577] env[62385]: DEBUG nova.virt.hardware [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 695.744745] env[62385]: DEBUG nova.virt.hardware [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 695.745883] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-783a5b6c-b7b7-4f76-a83e-00c52996382b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.753629] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-140c3fe5-4bc7-49a1-8595-f80d409943fe {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.766722] env[62385]: ERROR nova.compute.manager [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ad28927c-3e0a-47c2-b304-43949d2afc0d, please check neutron logs for more information. [ 695.766722] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Traceback (most recent call last): [ 695.766722] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 695.766722] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] yield resources [ 695.766722] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 695.766722] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] self.driver.spawn(context, instance, image_meta, [ 695.766722] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 695.766722] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] self._vmops.spawn(context, instance, image_meta, injected_files, [ 695.766722] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 695.766722] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] vm_ref = self.build_virtual_machine(instance, [ 695.766722] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 695.767167] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] vif_infos = vmwarevif.get_vif_info(self._session, [ 695.767167] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 695.767167] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] for vif in network_info: [ 695.767167] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 695.767167] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] return self._sync_wrapper(fn, *args, **kwargs) [ 695.767167] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 695.767167] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] self.wait() [ 695.767167] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 695.767167] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] self[:] = self._gt.wait() [ 695.767167] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 695.767167] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] return self._exit_event.wait() [ 695.767167] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 695.767167] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] current.throw(*self._exc) [ 695.767667] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 695.767667] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] result = function(*args, **kwargs) [ 695.767667] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 695.767667] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] return func(*args, **kwargs) [ 695.767667] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 695.767667] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] raise e [ 695.767667] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 695.767667] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] nwinfo = self.network_api.allocate_for_instance( [ 695.767667] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 695.767667] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] created_port_ids = self._update_ports_for_instance( [ 695.767667] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 695.767667] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] with excutils.save_and_reraise_exception(): [ 695.767667] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 695.768085] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] self.force_reraise() [ 695.768085] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 695.768085] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] raise self.value [ 695.768085] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 695.768085] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] updated_port = self._update_port( [ 695.768085] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 695.768085] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] _ensure_no_port_binding_failure(port) [ 695.768085] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 695.768085] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] raise exception.PortBindingFailed(port_id=port['id']) [ 695.768085] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] nova.exception.PortBindingFailed: Binding failed for port ad28927c-3e0a-47c2-b304-43949d2afc0d, please check neutron logs for more information. [ 695.768085] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] [ 695.768085] env[62385]: INFO nova.compute.manager [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Terminating instance [ 695.769349] env[62385]: DEBUG oslo_concurrency.lockutils [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] Acquiring lock "refresh_cache-4c96b9c6-40b9-4dea-8dd3-5e8e6372c261" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 695.890300] env[62385]: DEBUG nova.network.neutron [req-a15187f8-1588-4024-bf64-d4532f498a15 req-df2c2d2c-83cb-4598-985f-2aa6c662b126 service nova] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 695.905173] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.975897] env[62385]: DEBUG nova.network.neutron [req-a15187f8-1588-4024-bf64-d4532f498a15 req-df2c2d2c-83cb-4598-985f-2aa6c662b126 service nova] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.203582] env[62385]: DEBUG oslo_concurrency.lockutils [None req-50e22468-1913-42eb-9f13-0da6a2f22848 tempest-ServerShowV247Test-1343836024 tempest-ServerShowV247Test-1343836024-project-member] Lock "0ba32627-c646-477d-9062-b84624d01e48" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.187s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 696.476178] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5b07cdb-7572-4f15-946e-732aa0301400 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.483336] env[62385]: DEBUG oslo_concurrency.lockutils [req-a15187f8-1588-4024-bf64-d4532f498a15 req-df2c2d2c-83cb-4598-985f-2aa6c662b126 service nova] Releasing lock "refresh_cache-4c96b9c6-40b9-4dea-8dd3-5e8e6372c261" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.483829] env[62385]: DEBUG oslo_concurrency.lockutils [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] Acquired lock "refresh_cache-4c96b9c6-40b9-4dea-8dd3-5e8e6372c261" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.484032] env[62385]: DEBUG nova.network.neutron [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 696.485823] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a014ef1-17c4-47fa-a966-490ff70ad51f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.516072] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1873ba6-2c96-4d74-8e50-b7273998d857 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.523461] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4a3450c-9baa-40fa-b438-0057e17cf18b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.538257] env[62385]: DEBUG nova.compute.provider_tree [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 696.561920] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Acquiring lock "fd7a77e2-6d76-4d91-bdef-e30333247aa9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 696.562167] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Lock "fd7a77e2-6d76-4d91-bdef-e30333247aa9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 697.008844] env[62385]: DEBUG nova.network.neutron [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 697.041208] env[62385]: DEBUG nova.scheduler.client.report [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 697.104730] env[62385]: DEBUG nova.network.neutron [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.395817] env[62385]: DEBUG nova.compute.manager [req-e55559ed-a496-4a67-86bf-dce26408b297 req-7a8bbff9-ced8-4871-a76a-2b7bfad7be4c service nova] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Received event network-vif-deleted-ad28927c-3e0a-47c2-b304-43949d2afc0d {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 697.546466] env[62385]: DEBUG oslo_concurrency.lockutils [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.874s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 697.547230] env[62385]: ERROR nova.compute.manager [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port cb327ab2-1ebf-4a2a-9aec-b41148bf5183, please check neutron logs for more information. [ 697.547230] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Traceback (most recent call last): [ 697.547230] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 697.547230] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] self.driver.spawn(context, instance, image_meta, [ 697.547230] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 697.547230] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] self._vmops.spawn(context, instance, image_meta, injected_files, [ 697.547230] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 697.547230] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] vm_ref = self.build_virtual_machine(instance, [ 697.547230] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 697.547230] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] vif_infos = vmwarevif.get_vif_info(self._session, [ 697.547230] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 697.547600] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] for vif in network_info: [ 697.547600] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 697.547600] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] return self._sync_wrapper(fn, *args, **kwargs) [ 697.547600] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 697.547600] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] self.wait() [ 697.547600] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 697.547600] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] self[:] = self._gt.wait() [ 697.547600] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 697.547600] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] return self._exit_event.wait() [ 697.547600] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 697.547600] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] current.throw(*self._exc) [ 697.547600] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 697.547600] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] result = function(*args, **kwargs) [ 697.547947] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 697.547947] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] return func(*args, **kwargs) [ 697.547947] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 697.547947] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] raise e [ 697.547947] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 697.547947] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] nwinfo = self.network_api.allocate_for_instance( [ 697.547947] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 697.547947] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] created_port_ids = self._update_ports_for_instance( [ 697.547947] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 697.547947] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] with excutils.save_and_reraise_exception(): [ 697.547947] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 697.547947] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] self.force_reraise() [ 697.547947] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 697.548273] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] raise self.value [ 697.548273] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 697.548273] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] updated_port = self._update_port( [ 697.548273] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 697.548273] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] _ensure_no_port_binding_failure(port) [ 697.548273] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 697.548273] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] raise exception.PortBindingFailed(port_id=port['id']) [ 697.548273] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] nova.exception.PortBindingFailed: Binding failed for port cb327ab2-1ebf-4a2a-9aec-b41148bf5183, please check neutron logs for more information. [ 697.548273] env[62385]: ERROR nova.compute.manager [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] [ 697.548587] env[62385]: DEBUG nova.compute.utils [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Binding failed for port cb327ab2-1ebf-4a2a-9aec-b41148bf5183, please check neutron logs for more information. {{(pid=62385) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 697.549824] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.956s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 697.551397] env[62385]: INFO nova.compute.claims [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 697.554033] env[62385]: DEBUG nova.compute.manager [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Build of instance ea0b59fe-8dd9-4893-87ff-54706eb64479 was re-scheduled: Binding failed for port cb327ab2-1ebf-4a2a-9aec-b41148bf5183, please check neutron logs for more information. {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 697.554503] env[62385]: DEBUG nova.compute.manager [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Unplugging VIFs for instance {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 697.554759] env[62385]: DEBUG oslo_concurrency.lockutils [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] Acquiring lock "refresh_cache-ea0b59fe-8dd9-4893-87ff-54706eb64479" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 697.554929] env[62385]: DEBUG oslo_concurrency.lockutils [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] Acquired lock "refresh_cache-ea0b59fe-8dd9-4893-87ff-54706eb64479" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 697.555122] env[62385]: DEBUG nova.network.neutron [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 697.607842] env[62385]: DEBUG oslo_concurrency.lockutils [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] Releasing lock "refresh_cache-4c96b9c6-40b9-4dea-8dd3-5e8e6372c261" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 697.608058] env[62385]: DEBUG nova.compute.manager [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 697.608258] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 697.608549] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5446c519-2ec4-4de2-81cf-c133bf7b49d9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.619031] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7cc0dc3-8605-47c7-ba74-9e2d4bc9ae09 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.640168] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261 could not be found. [ 697.640380] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 697.640574] env[62385]: INFO nova.compute.manager [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Took 0.03 seconds to destroy the instance on the hypervisor. [ 697.640826] env[62385]: DEBUG oslo.service.loopingcall [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 697.641048] env[62385]: DEBUG nova.compute.manager [-] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 697.641143] env[62385]: DEBUG nova.network.neutron [-] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 697.658288] env[62385]: DEBUG nova.network.neutron [-] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 698.080990] env[62385]: DEBUG nova.network.neutron [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 698.161227] env[62385]: DEBUG nova.network.neutron [-] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.191362] env[62385]: DEBUG nova.network.neutron [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.665154] env[62385]: INFO nova.compute.manager [-] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Took 1.02 seconds to deallocate network for instance. [ 698.667965] env[62385]: DEBUG nova.compute.claims [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Aborting claim: {{(pid=62385) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 698.668448] env[62385]: DEBUG oslo_concurrency.lockutils [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 698.693272] env[62385]: DEBUG oslo_concurrency.lockutils [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] Releasing lock "refresh_cache-ea0b59fe-8dd9-4893-87ff-54706eb64479" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 698.693658] env[62385]: DEBUG nova.compute.manager [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 698.693863] env[62385]: DEBUG nova.compute.manager [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 698.694040] env[62385]: DEBUG nova.network.neutron [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 698.714125] env[62385]: DEBUG nova.network.neutron [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 698.972074] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aa24534-204b-4ec7-95e4-1b99230ca89e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.980203] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e62dce7a-5de9-42f8-8b31-5b593600b373 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.016322] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5669710f-3511-4723-aeb7-2d9f9bed5e5a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.025634] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dc8c886-ef56-466d-8e0b-3ad9a5613445 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.041526] env[62385]: DEBUG nova.compute.provider_tree [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 699.216468] env[62385]: DEBUG nova.network.neutron [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.544872] env[62385]: DEBUG nova.scheduler.client.report [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 699.722521] env[62385]: INFO nova.compute.manager [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] [instance: ea0b59fe-8dd9-4893-87ff-54706eb64479] Took 1.03 seconds to deallocate network for instance. [ 700.049763] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.500s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 700.050333] env[62385]: DEBUG nova.compute.manager [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 700.053391] env[62385]: DEBUG oslo_concurrency.lockutils [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.944s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 700.558773] env[62385]: DEBUG nova.compute.utils [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 700.562762] env[62385]: DEBUG nova.compute.manager [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 700.562929] env[62385]: DEBUG nova.network.neutron [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 700.613705] env[62385]: DEBUG nova.policy [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '12f9521f9c9b4c6b9afc4973be32c5b7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '731b6700ce5b43369b40165545f04950', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 700.762752] env[62385]: INFO nova.scheduler.client.report [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] Deleted allocations for instance ea0b59fe-8dd9-4893-87ff-54706eb64479 [ 700.913804] env[62385]: DEBUG nova.network.neutron [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Successfully created port: 7e3b5847-c00b-4e68-b344-f4e90204abf6 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 700.934917] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b435ab3-4032-4b9d-92e1-e36c8d48a2de {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.943529] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-045d39db-0856-4109-9003-afc463940a96 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.975842] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fc3cf62-129c-4a12-9541-4970b5cc8a64 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.983025] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-828c7f69-9264-49cf-b4ca-4afa65fbe801 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.997941] env[62385]: DEBUG nova.compute.provider_tree [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 701.066083] env[62385]: DEBUG nova.compute.manager [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 701.277337] env[62385]: DEBUG oslo_concurrency.lockutils [None req-004c46c5-78dc-47f6-9a8a-dfce6dc95a30 tempest-ImagesNegativeTestJSON-446811681 tempest-ImagesNegativeTestJSON-446811681-project-member] Lock "ea0b59fe-8dd9-4893-87ff-54706eb64479" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 139.105s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.348512] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Acquiring lock "d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 701.348819] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.502442] env[62385]: DEBUG nova.scheduler.client.report [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 701.616449] env[62385]: DEBUG nova.compute.manager [req-39ffc5f9-8f01-40f1-89af-bf8f0193ea92 req-403c213d-2805-41dd-8fee-23cb0bd71ace service nova] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Received event network-changed-7e3b5847-c00b-4e68-b344-f4e90204abf6 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 701.616695] env[62385]: DEBUG nova.compute.manager [req-39ffc5f9-8f01-40f1-89af-bf8f0193ea92 req-403c213d-2805-41dd-8fee-23cb0bd71ace service nova] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Refreshing instance network info cache due to event network-changed-7e3b5847-c00b-4e68-b344-f4e90204abf6. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 701.616858] env[62385]: DEBUG oslo_concurrency.lockutils [req-39ffc5f9-8f01-40f1-89af-bf8f0193ea92 req-403c213d-2805-41dd-8fee-23cb0bd71ace service nova] Acquiring lock "refresh_cache-866451cc-96c5-433a-a903-9faa1eed538a" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.616995] env[62385]: DEBUG oslo_concurrency.lockutils [req-39ffc5f9-8f01-40f1-89af-bf8f0193ea92 req-403c213d-2805-41dd-8fee-23cb0bd71ace service nova] Acquired lock "refresh_cache-866451cc-96c5-433a-a903-9faa1eed538a" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.617168] env[62385]: DEBUG nova.network.neutron [req-39ffc5f9-8f01-40f1-89af-bf8f0193ea92 req-403c213d-2805-41dd-8fee-23cb0bd71ace service nova] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Refreshing network info cache for port 7e3b5847-c00b-4e68-b344-f4e90204abf6 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 701.779167] env[62385]: DEBUG nova.compute.manager [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 701.794379] env[62385]: ERROR nova.compute.manager [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7e3b5847-c00b-4e68-b344-f4e90204abf6, please check neutron logs for more information. [ 701.794379] env[62385]: ERROR nova.compute.manager Traceback (most recent call last): [ 701.794379] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 701.794379] env[62385]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 701.794379] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 701.794379] env[62385]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 701.794379] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 701.794379] env[62385]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 701.794379] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.794379] env[62385]: ERROR nova.compute.manager self.force_reraise() [ 701.794379] env[62385]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.794379] env[62385]: ERROR nova.compute.manager raise self.value [ 701.794379] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 701.794379] env[62385]: ERROR nova.compute.manager updated_port = self._update_port( [ 701.794379] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.794379] env[62385]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 701.794841] env[62385]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.794841] env[62385]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 701.794841] env[62385]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7e3b5847-c00b-4e68-b344-f4e90204abf6, please check neutron logs for more information. [ 701.794841] env[62385]: ERROR nova.compute.manager [ 701.794841] env[62385]: Traceback (most recent call last): [ 701.794841] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 701.794841] env[62385]: listener.cb(fileno) [ 701.794841] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 701.794841] env[62385]: result = function(*args, **kwargs) [ 701.794841] env[62385]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 701.794841] env[62385]: return func(*args, **kwargs) [ 701.794841] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 701.794841] env[62385]: raise e [ 701.794841] env[62385]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 701.794841] env[62385]: nwinfo = self.network_api.allocate_for_instance( [ 701.794841] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 701.794841] env[62385]: created_port_ids = self._update_ports_for_instance( [ 701.794841] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 701.794841] env[62385]: with excutils.save_and_reraise_exception(): [ 701.794841] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.794841] env[62385]: self.force_reraise() [ 701.794841] env[62385]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.794841] env[62385]: raise self.value [ 701.794841] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 701.794841] env[62385]: updated_port = self._update_port( [ 701.794841] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.794841] env[62385]: _ensure_no_port_binding_failure(port) [ 701.794841] env[62385]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.794841] env[62385]: raise exception.PortBindingFailed(port_id=port['id']) [ 701.795778] env[62385]: nova.exception.PortBindingFailed: Binding failed for port 7e3b5847-c00b-4e68-b344-f4e90204abf6, please check neutron logs for more information. [ 701.795778] env[62385]: Removing descriptor: 17 [ 702.009456] env[62385]: DEBUG oslo_concurrency.lockutils [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.956s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 702.010090] env[62385]: ERROR nova.compute.manager [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0776ce7e-e90a-4326-ac85-6cdc70f2e785, please check neutron logs for more information. [ 702.010090] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Traceback (most recent call last): [ 702.010090] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 702.010090] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] self.driver.spawn(context, instance, image_meta, [ 702.010090] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 702.010090] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] self._vmops.spawn(context, instance, image_meta, injected_files, [ 702.010090] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 702.010090] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] vm_ref = self.build_virtual_machine(instance, [ 702.010090] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 702.010090] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] vif_infos = vmwarevif.get_vif_info(self._session, [ 702.010090] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 702.010573] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] for vif in network_info: [ 702.010573] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 702.010573] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] return self._sync_wrapper(fn, *args, **kwargs) [ 702.010573] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 702.010573] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] self.wait() [ 702.010573] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 702.010573] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] self[:] = self._gt.wait() [ 702.010573] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 702.010573] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] return self._exit_event.wait() [ 702.010573] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 702.010573] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] current.throw(*self._exc) [ 702.010573] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 702.010573] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] result = function(*args, **kwargs) [ 702.011188] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 702.011188] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] return func(*args, **kwargs) [ 702.011188] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 702.011188] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] raise e [ 702.011188] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 702.011188] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] nwinfo = self.network_api.allocate_for_instance( [ 702.011188] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 702.011188] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] created_port_ids = self._update_ports_for_instance( [ 702.011188] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 702.011188] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] with excutils.save_and_reraise_exception(): [ 702.011188] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 702.011188] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] self.force_reraise() [ 702.011188] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 702.012030] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] raise self.value [ 702.012030] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 702.012030] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] updated_port = self._update_port( [ 702.012030] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 702.012030] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] _ensure_no_port_binding_failure(port) [ 702.012030] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 702.012030] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] raise exception.PortBindingFailed(port_id=port['id']) [ 702.012030] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] nova.exception.PortBindingFailed: Binding failed for port 0776ce7e-e90a-4326-ac85-6cdc70f2e785, please check neutron logs for more information. [ 702.012030] env[62385]: ERROR nova.compute.manager [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] [ 702.012030] env[62385]: DEBUG nova.compute.utils [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Binding failed for port 0776ce7e-e90a-4326-ac85-6cdc70f2e785, please check neutron logs for more information. {{(pid=62385) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 702.012571] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.967s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 702.014055] env[62385]: INFO nova.compute.claims [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 702.018253] env[62385]: DEBUG nova.compute.manager [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Build of instance 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427 was re-scheduled: Binding failed for port 0776ce7e-e90a-4326-ac85-6cdc70f2e785, please check neutron logs for more information. {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 702.018253] env[62385]: DEBUG nova.compute.manager [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Unplugging VIFs for instance {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 702.018253] env[62385]: DEBUG oslo_concurrency.lockutils [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] Acquiring lock "refresh_cache-7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 702.018253] env[62385]: DEBUG oslo_concurrency.lockutils [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] Acquired lock "refresh_cache-7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.018541] env[62385]: DEBUG nova.network.neutron [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 702.075717] env[62385]: DEBUG nova.compute.manager [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 702.121382] env[62385]: DEBUG nova.virt.hardware [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 702.121619] env[62385]: DEBUG nova.virt.hardware [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 702.121779] env[62385]: DEBUG nova.virt.hardware [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 702.121953] env[62385]: DEBUG nova.virt.hardware [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 702.122122] env[62385]: DEBUG nova.virt.hardware [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 702.122278] env[62385]: DEBUG nova.virt.hardware [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 702.122486] env[62385]: DEBUG nova.virt.hardware [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 702.122647] env[62385]: DEBUG nova.virt.hardware [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 702.123117] env[62385]: DEBUG nova.virt.hardware [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 702.123117] env[62385]: DEBUG nova.virt.hardware [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 702.123192] env[62385]: DEBUG nova.virt.hardware [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 702.126316] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f22ec13b-2ffc-4224-8b35-e662a4c0ef93 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.136087] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00e1ed1f-87da-4ea1-9633-0dd26805b3b7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.150231] env[62385]: ERROR nova.compute.manager [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7e3b5847-c00b-4e68-b344-f4e90204abf6, please check neutron logs for more information. [ 702.150231] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Traceback (most recent call last): [ 702.150231] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 702.150231] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] yield resources [ 702.150231] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 702.150231] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] self.driver.spawn(context, instance, image_meta, [ 702.150231] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 702.150231] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 702.150231] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 702.150231] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] vm_ref = self.build_virtual_machine(instance, [ 702.150231] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 702.150821] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] vif_infos = vmwarevif.get_vif_info(self._session, [ 702.150821] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 702.150821] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] for vif in network_info: [ 702.150821] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 702.150821] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] return self._sync_wrapper(fn, *args, **kwargs) [ 702.150821] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 702.150821] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] self.wait() [ 702.150821] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 702.150821] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] self[:] = self._gt.wait() [ 702.150821] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 702.150821] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] return self._exit_event.wait() [ 702.150821] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 702.150821] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] current.throw(*self._exc) [ 702.151454] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 702.151454] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] result = function(*args, **kwargs) [ 702.151454] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 702.151454] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] return func(*args, **kwargs) [ 702.151454] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 702.151454] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] raise e [ 702.151454] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 702.151454] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] nwinfo = self.network_api.allocate_for_instance( [ 702.151454] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 702.151454] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] created_port_ids = self._update_ports_for_instance( [ 702.151454] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 702.151454] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] with excutils.save_and_reraise_exception(): [ 702.151454] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 702.151849] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] self.force_reraise() [ 702.151849] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 702.151849] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] raise self.value [ 702.151849] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 702.151849] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] updated_port = self._update_port( [ 702.151849] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 702.151849] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] _ensure_no_port_binding_failure(port) [ 702.151849] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 702.151849] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] raise exception.PortBindingFailed(port_id=port['id']) [ 702.151849] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] nova.exception.PortBindingFailed: Binding failed for port 7e3b5847-c00b-4e68-b344-f4e90204abf6, please check neutron logs for more information. [ 702.151849] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] [ 702.151849] env[62385]: INFO nova.compute.manager [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Terminating instance [ 702.153131] env[62385]: DEBUG nova.network.neutron [req-39ffc5f9-8f01-40f1-89af-bf8f0193ea92 req-403c213d-2805-41dd-8fee-23cb0bd71ace service nova] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 702.154986] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquiring lock "refresh_cache-866451cc-96c5-433a-a903-9faa1eed538a" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 702.236982] env[62385]: DEBUG nova.network.neutron [req-39ffc5f9-8f01-40f1-89af-bf8f0193ea92 req-403c213d-2805-41dd-8fee-23cb0bd71ace service nova] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.302456] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.535172] env[62385]: DEBUG nova.network.neutron [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 702.611683] env[62385]: DEBUG nova.network.neutron [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.739461] env[62385]: DEBUG oslo_concurrency.lockutils [req-39ffc5f9-8f01-40f1-89af-bf8f0193ea92 req-403c213d-2805-41dd-8fee-23cb0bd71ace service nova] Releasing lock "refresh_cache-866451cc-96c5-433a-a903-9faa1eed538a" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 702.739937] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquired lock "refresh_cache-866451cc-96c5-433a-a903-9faa1eed538a" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.740082] env[62385]: DEBUG nova.network.neutron [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 703.114082] env[62385]: DEBUG oslo_concurrency.lockutils [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] Releasing lock "refresh_cache-7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 703.114369] env[62385]: DEBUG nova.compute.manager [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 703.114506] env[62385]: DEBUG nova.compute.manager [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 703.114671] env[62385]: DEBUG nova.network.neutron [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 703.131881] env[62385]: DEBUG nova.network.neutron [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 703.264967] env[62385]: DEBUG nova.network.neutron [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 703.360572] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c847196-451f-4766-a3dd-a9c2a6395a50 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.369307] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49b6a852-317a-4823-a00b-2ef575b76372 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.380408] env[62385]: DEBUG nova.network.neutron [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.405469] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Releasing lock "refresh_cache-866451cc-96c5-433a-a903-9faa1eed538a" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 703.405886] env[62385]: DEBUG nova.compute.manager [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 703.406090] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 703.407138] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59719b7e-8900-493b-93a0-add7f123ad58 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.409725] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2d20b787-144d-435c-bf7c-7f2969c7aa44 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.417861] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67e5619a-2c63-47bc-9f70-72bfe5abd673 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.424558] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e135c43-b213-418f-8afa-b319b4175531 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.445478] env[62385]: DEBUG nova.compute.provider_tree [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 703.450666] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 866451cc-96c5-433a-a903-9faa1eed538a could not be found. [ 703.450666] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 703.450666] env[62385]: INFO nova.compute.manager [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 703.450919] env[62385]: DEBUG oslo.service.loopingcall [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 703.451293] env[62385]: DEBUG nova.compute.manager [-] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 703.451423] env[62385]: DEBUG nova.network.neutron [-] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 703.474076] env[62385]: DEBUG nova.network.neutron [-] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 703.637664] env[62385]: DEBUG nova.network.neutron [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.645627] env[62385]: DEBUG nova.compute.manager [req-62616f6d-db28-485c-9658-8911571f5d84 req-e80e42f5-775e-4576-9972-a7a3dc6945d9 service nova] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Received event network-vif-deleted-7e3b5847-c00b-4e68-b344-f4e90204abf6 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 703.952798] env[62385]: DEBUG nova.scheduler.client.report [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 703.977448] env[62385]: DEBUG nova.network.neutron [-] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.142373] env[62385]: INFO nova.compute.manager [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] [instance: 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427] Took 1.03 seconds to deallocate network for instance. [ 704.465021] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.450s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 704.465021] env[62385]: DEBUG nova.compute.manager [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 704.466394] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.442s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 704.467979] env[62385]: INFO nova.compute.claims [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 704.480947] env[62385]: INFO nova.compute.manager [-] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Took 1.03 seconds to deallocate network for instance. [ 704.481983] env[62385]: DEBUG nova.compute.claims [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Aborting claim: {{(pid=62385) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 704.481983] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.974449] env[62385]: DEBUG nova.compute.utils [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 704.978112] env[62385]: DEBUG nova.compute.manager [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 704.978112] env[62385]: DEBUG nova.network.neutron [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 705.024881] env[62385]: DEBUG nova.policy [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9badcc0f8b78496fb7df2c36b64a78f5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '72549cb052314f149111e4e24ffb6656', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 705.175920] env[62385]: INFO nova.scheduler.client.report [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] Deleted allocations for instance 7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427 [ 705.304362] env[62385]: DEBUG nova.network.neutron [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Successfully created port: 7965d15d-61f5-4fbb-8e38-53e0207a6c58 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 705.478772] env[62385]: DEBUG nova.compute.manager [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 705.687414] env[62385]: DEBUG oslo_concurrency.lockutils [None req-486f8d1e-c703-4430-bef7-f6ed048f29a3 tempest-InstanceActionsNegativeTestJSON-960405884 tempest-InstanceActionsNegativeTestJSON-960405884-project-member] Lock "7251e6e6-4cfe-444d-9dfb-ee0ad3a4c427" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 142.826s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.811023] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d99f737-24f9-421d-8048-c5d4f581da69 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.819534] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9508e004-6c65-4e49-bf50-b9c7d1e2ba14 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.850184] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4025f545-100f-4bdc-8478-6e25641cafe6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.857605] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a61a2012-ce5d-4381-96dc-e85c12334d96 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.870872] env[62385]: DEBUG nova.compute.provider_tree [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 706.192010] env[62385]: DEBUG nova.compute.manager [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 706.374307] env[62385]: DEBUG nova.scheduler.client.report [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 706.490939] env[62385]: DEBUG nova.compute.manager [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 706.519259] env[62385]: DEBUG nova.virt.hardware [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 706.519539] env[62385]: DEBUG nova.virt.hardware [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 706.519700] env[62385]: DEBUG nova.virt.hardware [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 706.519879] env[62385]: DEBUG nova.virt.hardware [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 706.520100] env[62385]: DEBUG nova.virt.hardware [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 706.520325] env[62385]: DEBUG nova.virt.hardware [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 706.520545] env[62385]: DEBUG nova.virt.hardware [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 706.520764] env[62385]: DEBUG nova.virt.hardware [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 706.520949] env[62385]: DEBUG nova.virt.hardware [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 706.521126] env[62385]: DEBUG nova.virt.hardware [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 706.521545] env[62385]: DEBUG nova.virt.hardware [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 706.522830] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae75c8a4-8693-4761-b796-65d6010c27c3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.532726] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7744b84e-7f6e-4753-943a-0b2cb1591305 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.717960] env[62385]: DEBUG oslo_concurrency.lockutils [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.864145] env[62385]: DEBUG nova.compute.manager [req-b64cd427-d9bb-4c9b-a90d-3cbce28b9f11 req-5a7740ff-2803-462c-864a-da759f71295d service nova] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Received event network-vif-plugged-7965d15d-61f5-4fbb-8e38-53e0207a6c58 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 706.864371] env[62385]: DEBUG oslo_concurrency.lockutils [req-b64cd427-d9bb-4c9b-a90d-3cbce28b9f11 req-5a7740ff-2803-462c-864a-da759f71295d service nova] Acquiring lock "1ddb4d6d-3872-49db-bb40-e21721241e89-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.864713] env[62385]: DEBUG oslo_concurrency.lockutils [req-b64cd427-d9bb-4c9b-a90d-3cbce28b9f11 req-5a7740ff-2803-462c-864a-da759f71295d service nova] Lock "1ddb4d6d-3872-49db-bb40-e21721241e89-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.864808] env[62385]: DEBUG oslo_concurrency.lockutils [req-b64cd427-d9bb-4c9b-a90d-3cbce28b9f11 req-5a7740ff-2803-462c-864a-da759f71295d service nova] Lock "1ddb4d6d-3872-49db-bb40-e21721241e89-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.864896] env[62385]: DEBUG nova.compute.manager [req-b64cd427-d9bb-4c9b-a90d-3cbce28b9f11 req-5a7740ff-2803-462c-864a-da759f71295d service nova] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] No waiting events found dispatching network-vif-plugged-7965d15d-61f5-4fbb-8e38-53e0207a6c58 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 706.865065] env[62385]: WARNING nova.compute.manager [req-b64cd427-d9bb-4c9b-a90d-3cbce28b9f11 req-5a7740ff-2803-462c-864a-da759f71295d service nova] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Received unexpected event network-vif-plugged-7965d15d-61f5-4fbb-8e38-53e0207a6c58 for instance with vm_state building and task_state spawning. [ 706.879891] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.413s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.880408] env[62385]: DEBUG nova.compute.manager [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 706.886023] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.827s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.886023] env[62385]: INFO nova.compute.claims [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 706.972598] env[62385]: DEBUG nova.network.neutron [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Successfully updated port: 7965d15d-61f5-4fbb-8e38-53e0207a6c58 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 707.392022] env[62385]: DEBUG nova.compute.utils [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 707.394030] env[62385]: DEBUG nova.compute.manager [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 707.394332] env[62385]: DEBUG nova.network.neutron [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 707.463744] env[62385]: DEBUG nova.policy [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9badcc0f8b78496fb7df2c36b64a78f5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '72549cb052314f149111e4e24ffb6656', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 707.480947] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Acquiring lock "refresh_cache-1ddb4d6d-3872-49db-bb40-e21721241e89" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.480947] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Acquired lock "refresh_cache-1ddb4d6d-3872-49db-bb40-e21721241e89" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.480947] env[62385]: DEBUG nova.network.neutron [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 707.894748] env[62385]: DEBUG nova.compute.manager [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 708.042591] env[62385]: DEBUG nova.network.neutron [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 708.163215] env[62385]: DEBUG nova.network.neutron [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Successfully created port: 402f6138-a737-4b40-acf1-3776bf4db1e5 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 708.332425] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2de6d92-5cdf-45af-a386-f2b2f4c66222 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.335814] env[62385]: DEBUG nova.network.neutron [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Updating instance_info_cache with network_info: [{"id": "7965d15d-61f5-4fbb-8e38-53e0207a6c58", "address": "fa:16:3e:fb:f0:ae", "network": {"id": "1d0e9098-63d3-4df5-a80a-06f9a5d90db6", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-303269893-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72549cb052314f149111e4e24ffb6656", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "572b7281-aad3-45fa-9cb2-fc1c70569948", "external-id": "nsx-vlan-transportzone-722", "segmentation_id": 722, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7965d15d-61", "ovs_interfaceid": "7965d15d-61f5-4fbb-8e38-53e0207a6c58", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.342280] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fb3a639-3fef-4109-b87e-dfd8aafb626a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.381203] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e93920bb-d382-45e8-b6a3-a4b58db7bef4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.391751] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6957486-27c1-40aa-aee5-aaaa91741c56 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.410098] env[62385]: DEBUG nova.compute.provider_tree [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 708.838276] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Releasing lock "refresh_cache-1ddb4d6d-3872-49db-bb40-e21721241e89" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.838693] env[62385]: DEBUG nova.compute.manager [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Instance network_info: |[{"id": "7965d15d-61f5-4fbb-8e38-53e0207a6c58", "address": "fa:16:3e:fb:f0:ae", "network": {"id": "1d0e9098-63d3-4df5-a80a-06f9a5d90db6", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-303269893-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72549cb052314f149111e4e24ffb6656", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "572b7281-aad3-45fa-9cb2-fc1c70569948", "external-id": "nsx-vlan-transportzone-722", "segmentation_id": 722, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7965d15d-61", "ovs_interfaceid": "7965d15d-61f5-4fbb-8e38-53e0207a6c58", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 708.839196] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fb:f0:ae', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '572b7281-aad3-45fa-9cb2-fc1c70569948', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7965d15d-61f5-4fbb-8e38-53e0207a6c58', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 708.848093] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Creating folder: Project (72549cb052314f149111e4e24ffb6656). Parent ref: group-v261107. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 708.848093] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4f70905a-a906-46ae-bb15-df5df9619a0b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.861961] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Created folder: Project (72549cb052314f149111e4e24ffb6656) in parent group-v261107. [ 708.861961] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Creating folder: Instances. Parent ref: group-v261120. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 708.861961] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9b7ccb6c-dad1-458d-b5b6-0341d526cc90 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.871703] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Created folder: Instances in parent group-v261120. [ 708.871948] env[62385]: DEBUG oslo.service.loopingcall [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 708.872159] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 708.872367] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7b5952e5-847d-490f-b262-cabf959d59f9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.891314] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 708.891314] env[62385]: value = "task-1205484" [ 708.891314] env[62385]: _type = "Task" [ 708.891314] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.900555] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205484, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.903381] env[62385]: DEBUG nova.compute.manager [req-202e32fc-ef0e-48e8-a7c9-618a04dfd6ec req-91f31e49-178d-440a-a28b-58cc129fdf25 service nova] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Received event network-changed-7965d15d-61f5-4fbb-8e38-53e0207a6c58 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 708.903729] env[62385]: DEBUG nova.compute.manager [req-202e32fc-ef0e-48e8-a7c9-618a04dfd6ec req-91f31e49-178d-440a-a28b-58cc129fdf25 service nova] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Refreshing instance network info cache due to event network-changed-7965d15d-61f5-4fbb-8e38-53e0207a6c58. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 708.907351] env[62385]: DEBUG oslo_concurrency.lockutils [req-202e32fc-ef0e-48e8-a7c9-618a04dfd6ec req-91f31e49-178d-440a-a28b-58cc129fdf25 service nova] Acquiring lock "refresh_cache-1ddb4d6d-3872-49db-bb40-e21721241e89" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 708.907351] env[62385]: DEBUG oslo_concurrency.lockutils [req-202e32fc-ef0e-48e8-a7c9-618a04dfd6ec req-91f31e49-178d-440a-a28b-58cc129fdf25 service nova] Acquired lock "refresh_cache-1ddb4d6d-3872-49db-bb40-e21721241e89" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.907351] env[62385]: DEBUG nova.network.neutron [req-202e32fc-ef0e-48e8-a7c9-618a04dfd6ec req-91f31e49-178d-440a-a28b-58cc129fdf25 service nova] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Refreshing network info cache for port 7965d15d-61f5-4fbb-8e38-53e0207a6c58 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 708.917596] env[62385]: DEBUG nova.compute.manager [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 708.919120] env[62385]: DEBUG nova.scheduler.client.report [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 708.959153] env[62385]: DEBUG nova.virt.hardware [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 708.959153] env[62385]: DEBUG nova.virt.hardware [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 708.959153] env[62385]: DEBUG nova.virt.hardware [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 708.959342] env[62385]: DEBUG nova.virt.hardware [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 708.959342] env[62385]: DEBUG nova.virt.hardware [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 708.959342] env[62385]: DEBUG nova.virt.hardware [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 708.959503] env[62385]: DEBUG nova.virt.hardware [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 708.959806] env[62385]: DEBUG nova.virt.hardware [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 708.960116] env[62385]: DEBUG nova.virt.hardware [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 708.960415] env[62385]: DEBUG nova.virt.hardware [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 708.960777] env[62385]: DEBUG nova.virt.hardware [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 708.962131] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bdb5042-b74e-4a1b-9daa-518a799d0245 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.973078] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6aa59601-5579-4520-8b87-89122ee57f6f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.401406] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205484, 'name': CreateVM_Task, 'duration_secs': 0.3339} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.401583] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 709.412117] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 709.412117] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.412117] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 709.412117] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a69f24da-4cf2-49f8-b737-d9a9e62c6a04 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.416842] env[62385]: DEBUG oslo_vmware.api [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Waiting for the task: (returnval){ [ 709.416842] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c076f8-545a-2012-d49c-da7c4c7bfd2d" [ 709.416842] env[62385]: _type = "Task" [ 709.416842] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.424875] env[62385]: DEBUG oslo_vmware.api [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c076f8-545a-2012-d49c-da7c4c7bfd2d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.426691] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.543s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.427151] env[62385]: DEBUG nova.compute.manager [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 709.430276] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.953s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 709.711348] env[62385]: DEBUG nova.network.neutron [req-202e32fc-ef0e-48e8-a7c9-618a04dfd6ec req-91f31e49-178d-440a-a28b-58cc129fdf25 service nova] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Updated VIF entry in instance network info cache for port 7965d15d-61f5-4fbb-8e38-53e0207a6c58. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 709.711796] env[62385]: DEBUG nova.network.neutron [req-202e32fc-ef0e-48e8-a7c9-618a04dfd6ec req-91f31e49-178d-440a-a28b-58cc129fdf25 service nova] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Updating instance_info_cache with network_info: [{"id": "7965d15d-61f5-4fbb-8e38-53e0207a6c58", "address": "fa:16:3e:fb:f0:ae", "network": {"id": "1d0e9098-63d3-4df5-a80a-06f9a5d90db6", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-303269893-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72549cb052314f149111e4e24ffb6656", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "572b7281-aad3-45fa-9cb2-fc1c70569948", "external-id": "nsx-vlan-transportzone-722", "segmentation_id": 722, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7965d15d-61", "ovs_interfaceid": "7965d15d-61f5-4fbb-8e38-53e0207a6c58", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.905336] env[62385]: DEBUG nova.network.neutron [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Successfully updated port: 402f6138-a737-4b40-acf1-3776bf4db1e5 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 709.929599] env[62385]: DEBUG oslo_vmware.api [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c076f8-545a-2012-d49c-da7c4c7bfd2d, 'name': SearchDatastore_Task, 'duration_secs': 0.012013} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.929599] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 709.929783] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 709.930022] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 709.930171] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.930340] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 709.931034] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cdadabc3-c646-49cb-b917-557a50012bc8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.936740] env[62385]: DEBUG nova.compute.utils [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 709.940841] env[62385]: DEBUG nova.compute.manager [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 709.941012] env[62385]: DEBUG nova.network.neutron [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 709.942917] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 709.943091] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 709.943836] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d1414c21-acd1-44f6-b24a-cdd9670f8ec8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.952502] env[62385]: DEBUG oslo_vmware.api [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Waiting for the task: (returnval){ [ 709.952502] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a83359-4a3b-87c5-4326-0ec054514554" [ 709.952502] env[62385]: _type = "Task" [ 709.952502] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.964024] env[62385]: DEBUG oslo_vmware.api [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a83359-4a3b-87c5-4326-0ec054514554, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.987448] env[62385]: DEBUG nova.policy [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9badcc0f8b78496fb7df2c36b64a78f5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '72549cb052314f149111e4e24ffb6656', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 710.218527] env[62385]: DEBUG oslo_concurrency.lockutils [req-202e32fc-ef0e-48e8-a7c9-618a04dfd6ec req-91f31e49-178d-440a-a28b-58cc129fdf25 service nova] Releasing lock "refresh_cache-1ddb4d6d-3872-49db-bb40-e21721241e89" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.279674] env[62385]: DEBUG nova.network.neutron [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Successfully created port: 60df554a-a3e2-41eb-b893-16459c19bc7c {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 710.384420] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20189cc0-de1a-4275-85fb-25e04eb48d05 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.392014] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60de98e2-027d-41b1-8c99-f8152debc68a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.424470] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Acquiring lock "refresh_cache-4a4b254c-46a3-4db6-9c0f-f476ae8f2b69" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.424470] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Acquired lock "refresh_cache-4a4b254c-46a3-4db6-9c0f-f476ae8f2b69" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.424470] env[62385]: DEBUG nova.network.neutron [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 710.430144] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d73e9f8-56bf-4f7c-b27d-bb5a87051d42 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.441399] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dfb7cd0-30ae-4bfe-ab3d-dfcb59b82411 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.446163] env[62385]: DEBUG nova.compute.manager [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 710.462164] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 710.462933] env[62385]: DEBUG nova.compute.provider_tree [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 710.468144] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 710.476960] env[62385]: DEBUG oslo_vmware.api [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a83359-4a3b-87c5-4326-0ec054514554, 'name': SearchDatastore_Task, 'duration_secs': 0.020827} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.478325] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-82eac850-db9c-4fdf-98ec-f357e977347f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.484136] env[62385]: DEBUG oslo_vmware.api [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Waiting for the task: (returnval){ [ 710.484136] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52d9cc35-31be-7ac7-73aa-9889d8f31af1" [ 710.484136] env[62385]: _type = "Task" [ 710.484136] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.493871] env[62385]: DEBUG oslo_vmware.api [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52d9cc35-31be-7ac7-73aa-9889d8f31af1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.943156] env[62385]: DEBUG nova.compute.manager [req-9b6981b4-6983-4fe0-8627-963ce878999e req-5ec560d4-f010-46b7-be06-010c88e63d75 service nova] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Received event network-vif-plugged-402f6138-a737-4b40-acf1-3776bf4db1e5 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 710.943156] env[62385]: DEBUG oslo_concurrency.lockutils [req-9b6981b4-6983-4fe0-8627-963ce878999e req-5ec560d4-f010-46b7-be06-010c88e63d75 service nova] Acquiring lock "4a4b254c-46a3-4db6-9c0f-f476ae8f2b69-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 710.943156] env[62385]: DEBUG oslo_concurrency.lockutils [req-9b6981b4-6983-4fe0-8627-963ce878999e req-5ec560d4-f010-46b7-be06-010c88e63d75 service nova] Lock "4a4b254c-46a3-4db6-9c0f-f476ae8f2b69-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.943156] env[62385]: DEBUG oslo_concurrency.lockutils [req-9b6981b4-6983-4fe0-8627-963ce878999e req-5ec560d4-f010-46b7-be06-010c88e63d75 service nova] Lock "4a4b254c-46a3-4db6-9c0f-f476ae8f2b69-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.943156] env[62385]: DEBUG nova.compute.manager [req-9b6981b4-6983-4fe0-8627-963ce878999e req-5ec560d4-f010-46b7-be06-010c88e63d75 service nova] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] No waiting events found dispatching network-vif-plugged-402f6138-a737-4b40-acf1-3776bf4db1e5 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 710.943632] env[62385]: WARNING nova.compute.manager [req-9b6981b4-6983-4fe0-8627-963ce878999e req-5ec560d4-f010-46b7-be06-010c88e63d75 service nova] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Received unexpected event network-vif-plugged-402f6138-a737-4b40-acf1-3776bf4db1e5 for instance with vm_state building and task_state spawning. [ 710.943632] env[62385]: DEBUG nova.compute.manager [req-9b6981b4-6983-4fe0-8627-963ce878999e req-5ec560d4-f010-46b7-be06-010c88e63d75 service nova] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Received event network-changed-402f6138-a737-4b40-acf1-3776bf4db1e5 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 710.944752] env[62385]: DEBUG nova.compute.manager [req-9b6981b4-6983-4fe0-8627-963ce878999e req-5ec560d4-f010-46b7-be06-010c88e63d75 service nova] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Refreshing instance network info cache due to event network-changed-402f6138-a737-4b40-acf1-3776bf4db1e5. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 710.945432] env[62385]: DEBUG oslo_concurrency.lockutils [req-9b6981b4-6983-4fe0-8627-963ce878999e req-5ec560d4-f010-46b7-be06-010c88e63d75 service nova] Acquiring lock "refresh_cache-4a4b254c-46a3-4db6-9c0f-f476ae8f2b69" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.970718] env[62385]: DEBUG nova.scheduler.client.report [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 710.982260] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 710.982685] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Starting heal instance info cache {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 710.986219] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Rebuilding the list of instances to heal {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 711.001020] env[62385]: DEBUG oslo_vmware.api [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52d9cc35-31be-7ac7-73aa-9889d8f31af1, 'name': SearchDatastore_Task, 'duration_secs': 0.009584} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.001020] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 711.001020] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 1ddb4d6d-3872-49db-bb40-e21721241e89/1ddb4d6d-3872-49db-bb40-e21721241e89.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 711.001020] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3094808e-98ed-4bff-a445-7a63711095b5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.009309] env[62385]: DEBUG oslo_vmware.api [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Waiting for the task: (returnval){ [ 711.009309] env[62385]: value = "task-1205485" [ 711.009309] env[62385]: _type = "Task" [ 711.009309] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.009309] env[62385]: DEBUG nova.network.neutron [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 711.020596] env[62385]: DEBUG oslo_vmware.api [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205485, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.398197] env[62385]: DEBUG nova.network.neutron [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Updating instance_info_cache with network_info: [{"id": "402f6138-a737-4b40-acf1-3776bf4db1e5", "address": "fa:16:3e:d5:1e:ac", "network": {"id": "1d0e9098-63d3-4df5-a80a-06f9a5d90db6", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-303269893-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72549cb052314f149111e4e24ffb6656", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "572b7281-aad3-45fa-9cb2-fc1c70569948", "external-id": "nsx-vlan-transportzone-722", "segmentation_id": 722, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap402f6138-a7", "ovs_interfaceid": "402f6138-a737-4b40-acf1-3776bf4db1e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.455870] env[62385]: DEBUG nova.compute.manager [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 711.478922] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.048s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.479610] env[62385]: ERROR nova.compute.manager [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 20df7395-1da3-486e-a35e-0279141f722a, please check neutron logs for more information. [ 711.479610] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Traceback (most recent call last): [ 711.479610] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 711.479610] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] self.driver.spawn(context, instance, image_meta, [ 711.479610] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 711.479610] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] self._vmops.spawn(context, instance, image_meta, injected_files, [ 711.479610] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 711.479610] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] vm_ref = self.build_virtual_machine(instance, [ 711.479610] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 711.479610] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] vif_infos = vmwarevif.get_vif_info(self._session, [ 711.479610] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 711.480030] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] for vif in network_info: [ 711.480030] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 711.480030] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] return self._sync_wrapper(fn, *args, **kwargs) [ 711.480030] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 711.480030] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] self.wait() [ 711.480030] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 711.480030] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] self[:] = self._gt.wait() [ 711.480030] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 711.480030] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] return self._exit_event.wait() [ 711.480030] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 711.480030] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] result = hub.switch() [ 711.480030] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 711.480030] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] return self.greenlet.switch() [ 711.480453] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 711.480453] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] result = function(*args, **kwargs) [ 711.480453] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 711.480453] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] return func(*args, **kwargs) [ 711.480453] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 711.480453] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] raise e [ 711.480453] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 711.480453] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] nwinfo = self.network_api.allocate_for_instance( [ 711.480453] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 711.480453] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] created_port_ids = self._update_ports_for_instance( [ 711.480453] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 711.480453] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] with excutils.save_and_reraise_exception(): [ 711.480453] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 711.480816] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] self.force_reraise() [ 711.480816] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 711.480816] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] raise self.value [ 711.480816] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 711.480816] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] updated_port = self._update_port( [ 711.480816] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 711.480816] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] _ensure_no_port_binding_failure(port) [ 711.480816] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 711.480816] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] raise exception.PortBindingFailed(port_id=port['id']) [ 711.480816] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] nova.exception.PortBindingFailed: Binding failed for port 20df7395-1da3-486e-a35e-0279141f722a, please check neutron logs for more information. [ 711.480816] env[62385]: ERROR nova.compute.manager [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] [ 711.484027] env[62385]: DEBUG nova.compute.utils [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Binding failed for port 20df7395-1da3-486e-a35e-0279141f722a, please check neutron logs for more information. {{(pid=62385) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 711.484027] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.426s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 711.487835] env[62385]: DEBUG nova.compute.manager [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Build of instance 1b459bac-e27d-4af5-bfc2-49898f5ced25 was re-scheduled: Binding failed for port 20df7395-1da3-486e-a35e-0279141f722a, please check neutron logs for more information. {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 711.487835] env[62385]: DEBUG nova.compute.manager [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Unplugging VIFs for instance {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 711.487835] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] Acquiring lock "refresh_cache-1b459bac-e27d-4af5-bfc2-49898f5ced25" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 711.487835] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] Acquired lock "refresh_cache-1b459bac-e27d-4af5-bfc2-49898f5ced25" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.488293] env[62385]: DEBUG nova.network.neutron [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 711.492263] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Skipping network cache update for instance because it is Building. {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 711.492406] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Skipping network cache update for instance because it is Building. {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 711.492531] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Skipping network cache update for instance because it is Building. {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 711.492655] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Skipping network cache update for instance because it is Building. {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 711.492837] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Skipping network cache update for instance because it is Building. {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 711.492899] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Skipping network cache update for instance because it is Building. {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 711.493022] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Didn't find any instances for network info cache update. {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 711.496621] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 711.496835] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 711.497047] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 711.497235] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 711.497399] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 711.497573] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 711.497710] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62385) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 711.497881] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager.update_available_resource {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 711.508225] env[62385]: DEBUG nova.virt.hardware [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:16:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 711.508470] env[62385]: DEBUG nova.virt.hardware [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 711.508632] env[62385]: DEBUG nova.virt.hardware [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 711.508878] env[62385]: DEBUG nova.virt.hardware [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 711.509089] env[62385]: DEBUG nova.virt.hardware [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 711.509255] env[62385]: DEBUG nova.virt.hardware [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 711.509466] env[62385]: DEBUG nova.virt.hardware [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 711.509623] env[62385]: DEBUG nova.virt.hardware [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 711.509791] env[62385]: DEBUG nova.virt.hardware [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 711.509951] env[62385]: DEBUG nova.virt.hardware [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 711.510142] env[62385]: DEBUG nova.virt.hardware [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 711.511728] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a92005bd-b51b-4d11-9acd-973426034d57 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.526668] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f61d469-8f1c-4dcd-901e-52c9eebc5976 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.530346] env[62385]: DEBUG oslo_vmware.api [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205485, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.477396} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.530873] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 1ddb4d6d-3872-49db-bb40-e21721241e89/1ddb4d6d-3872-49db-bb40-e21721241e89.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 711.531102] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 711.531711] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-410c2de2-8d9b-4e8a-b368-ec2e08ab362b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.549818] env[62385]: DEBUG oslo_vmware.api [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Waiting for the task: (returnval){ [ 711.549818] env[62385]: value = "task-1205486" [ 711.549818] env[62385]: _type = "Task" [ 711.549818] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.557960] env[62385]: DEBUG oslo_vmware.api [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205486, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.875418] env[62385]: DEBUG nova.network.neutron [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Successfully updated port: 60df554a-a3e2-41eb-b893-16459c19bc7c {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 711.899601] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Releasing lock "refresh_cache-4a4b254c-46a3-4db6-9c0f-f476ae8f2b69" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 711.899981] env[62385]: DEBUG nova.compute.manager [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Instance network_info: |[{"id": "402f6138-a737-4b40-acf1-3776bf4db1e5", "address": "fa:16:3e:d5:1e:ac", "network": {"id": "1d0e9098-63d3-4df5-a80a-06f9a5d90db6", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-303269893-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72549cb052314f149111e4e24ffb6656", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "572b7281-aad3-45fa-9cb2-fc1c70569948", "external-id": "nsx-vlan-transportzone-722", "segmentation_id": 722, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap402f6138-a7", "ovs_interfaceid": "402f6138-a737-4b40-acf1-3776bf4db1e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 711.900299] env[62385]: DEBUG oslo_concurrency.lockutils [req-9b6981b4-6983-4fe0-8627-963ce878999e req-5ec560d4-f010-46b7-be06-010c88e63d75 service nova] Acquired lock "refresh_cache-4a4b254c-46a3-4db6-9c0f-f476ae8f2b69" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.900475] env[62385]: DEBUG nova.network.neutron [req-9b6981b4-6983-4fe0-8627-963ce878999e req-5ec560d4-f010-46b7-be06-010c88e63d75 service nova] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Refreshing network info cache for port 402f6138-a737-4b40-acf1-3776bf4db1e5 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 711.901637] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d5:1e:ac', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '572b7281-aad3-45fa-9cb2-fc1c70569948', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '402f6138-a737-4b40-acf1-3776bf4db1e5', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 711.910953] env[62385]: DEBUG oslo.service.loopingcall [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 711.911667] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 711.912336] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0762154b-f4e9-4fda-acda-e52052be3823 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.934327] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 711.934327] env[62385]: value = "task-1205487" [ 711.934327] env[62385]: _type = "Task" [ 711.934327] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.942852] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205487, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.003318] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.018162] env[62385]: DEBUG nova.network.neutron [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 712.067923] env[62385]: DEBUG oslo_vmware.api [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205486, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072195} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.068274] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 712.069021] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99850e90-8b55-4d20-9bbc-460fb4f90a76 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.099041] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Reconfiguring VM instance instance-00000029 to attach disk [datastore1] 1ddb4d6d-3872-49db-bb40-e21721241e89/1ddb4d6d-3872-49db-bb40-e21721241e89.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 712.103585] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-daac11fa-eb01-4d21-8774-c3ff0337a59c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.123344] env[62385]: DEBUG oslo_vmware.api [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Waiting for the task: (returnval){ [ 712.123344] env[62385]: value = "task-1205488" [ 712.123344] env[62385]: _type = "Task" [ 712.123344] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.135506] env[62385]: DEBUG oslo_vmware.api [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205488, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.171634] env[62385]: DEBUG nova.network.neutron [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.328262] env[62385]: DEBUG oslo_concurrency.lockutils [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "9d6f098a-0b05-43ef-96b0-9eb99ad3538c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.328262] env[62385]: DEBUG oslo_concurrency.lockutils [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "9d6f098a-0b05-43ef-96b0-9eb99ad3538c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.378189] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Acquiring lock "refresh_cache-3e184125-28af-469b-83cc-4ab8859e0e2f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.378332] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Acquired lock "refresh_cache-3e184125-28af-469b-83cc-4ab8859e0e2f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.378483] env[62385]: DEBUG nova.network.neutron [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 712.420565] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa13a3a3-9ec2-4bfe-a20f-345b5d63e92a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.428481] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e96f9dcb-28fe-4461-92a0-e1fc184c9558 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.467063] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8dd5cf5-402c-4159-8ed5-9b4dac2677fa {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.478452] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d8a0a22-dc63-46fb-9386-1077c963db0f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.482250] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205487, 'name': CreateVM_Task, 'duration_secs': 0.35005} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.482344] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 712.483848] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.484013] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.484333] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 712.484568] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-df2dbc49-67d3-4e28-927f-8b8bc2fb2649 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.494955] env[62385]: DEBUG nova.compute.provider_tree [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 712.500138] env[62385]: DEBUG oslo_vmware.api [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Waiting for the task: (returnval){ [ 712.500138] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5289567e-a285-3fbc-8f03-30134cb84e47" [ 712.500138] env[62385]: _type = "Task" [ 712.500138] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.507317] env[62385]: DEBUG oslo_vmware.api [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5289567e-a285-3fbc-8f03-30134cb84e47, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.636169] env[62385]: DEBUG oslo_vmware.api [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205488, 'name': ReconfigVM_Task, 'duration_secs': 0.268902} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.636169] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Reconfigured VM instance instance-00000029 to attach disk [datastore1] 1ddb4d6d-3872-49db-bb40-e21721241e89/1ddb4d6d-3872-49db-bb40-e21721241e89.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 712.636662] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8b07cd81-43cf-4761-8f26-719d265de86b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.643431] env[62385]: DEBUG oslo_vmware.api [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Waiting for the task: (returnval){ [ 712.643431] env[62385]: value = "task-1205489" [ 712.643431] env[62385]: _type = "Task" [ 712.643431] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.651311] env[62385]: DEBUG oslo_vmware.api [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205489, 'name': Rename_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.675576] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] Releasing lock "refresh_cache-1b459bac-e27d-4af5-bfc2-49898f5ced25" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.675821] env[62385]: DEBUG nova.compute.manager [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 712.676015] env[62385]: DEBUG nova.compute.manager [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 712.676189] env[62385]: DEBUG nova.network.neutron [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 712.692896] env[62385]: DEBUG nova.network.neutron [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 712.857864] env[62385]: DEBUG nova.network.neutron [req-9b6981b4-6983-4fe0-8627-963ce878999e req-5ec560d4-f010-46b7-be06-010c88e63d75 service nova] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Updated VIF entry in instance network info cache for port 402f6138-a737-4b40-acf1-3776bf4db1e5. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 712.858229] env[62385]: DEBUG nova.network.neutron [req-9b6981b4-6983-4fe0-8627-963ce878999e req-5ec560d4-f010-46b7-be06-010c88e63d75 service nova] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Updating instance_info_cache with network_info: [{"id": "402f6138-a737-4b40-acf1-3776bf4db1e5", "address": "fa:16:3e:d5:1e:ac", "network": {"id": "1d0e9098-63d3-4df5-a80a-06f9a5d90db6", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-303269893-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72549cb052314f149111e4e24ffb6656", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "572b7281-aad3-45fa-9cb2-fc1c70569948", "external-id": "nsx-vlan-transportzone-722", "segmentation_id": 722, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap402f6138-a7", "ovs_interfaceid": "402f6138-a737-4b40-acf1-3776bf4db1e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.910480] env[62385]: DEBUG nova.network.neutron [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 712.980046] env[62385]: DEBUG nova.compute.manager [req-207d236e-b8c9-43d9-bfa4-d3d0f83ec11e req-2e6ee74b-6615-4153-b3fd-c1fdc598b530 service nova] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Received event network-vif-plugged-60df554a-a3e2-41eb-b893-16459c19bc7c {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 712.980046] env[62385]: DEBUG oslo_concurrency.lockutils [req-207d236e-b8c9-43d9-bfa4-d3d0f83ec11e req-2e6ee74b-6615-4153-b3fd-c1fdc598b530 service nova] Acquiring lock "3e184125-28af-469b-83cc-4ab8859e0e2f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.980149] env[62385]: DEBUG oslo_concurrency.lockutils [req-207d236e-b8c9-43d9-bfa4-d3d0f83ec11e req-2e6ee74b-6615-4153-b3fd-c1fdc598b530 service nova] Lock "3e184125-28af-469b-83cc-4ab8859e0e2f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.980413] env[62385]: DEBUG oslo_concurrency.lockutils [req-207d236e-b8c9-43d9-bfa4-d3d0f83ec11e req-2e6ee74b-6615-4153-b3fd-c1fdc598b530 service nova] Lock "3e184125-28af-469b-83cc-4ab8859e0e2f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.980489] env[62385]: DEBUG nova.compute.manager [req-207d236e-b8c9-43d9-bfa4-d3d0f83ec11e req-2e6ee74b-6615-4153-b3fd-c1fdc598b530 service nova] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] No waiting events found dispatching network-vif-plugged-60df554a-a3e2-41eb-b893-16459c19bc7c {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 712.980751] env[62385]: WARNING nova.compute.manager [req-207d236e-b8c9-43d9-bfa4-d3d0f83ec11e req-2e6ee74b-6615-4153-b3fd-c1fdc598b530 service nova] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Received unexpected event network-vif-plugged-60df554a-a3e2-41eb-b893-16459c19bc7c for instance with vm_state building and task_state spawning. [ 712.980869] env[62385]: DEBUG nova.compute.manager [req-207d236e-b8c9-43d9-bfa4-d3d0f83ec11e req-2e6ee74b-6615-4153-b3fd-c1fdc598b530 service nova] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Received event network-changed-60df554a-a3e2-41eb-b893-16459c19bc7c {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 712.981047] env[62385]: DEBUG nova.compute.manager [req-207d236e-b8c9-43d9-bfa4-d3d0f83ec11e req-2e6ee74b-6615-4153-b3fd-c1fdc598b530 service nova] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Refreshing instance network info cache due to event network-changed-60df554a-a3e2-41eb-b893-16459c19bc7c. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 712.981214] env[62385]: DEBUG oslo_concurrency.lockutils [req-207d236e-b8c9-43d9-bfa4-d3d0f83ec11e req-2e6ee74b-6615-4153-b3fd-c1fdc598b530 service nova] Acquiring lock "refresh_cache-3e184125-28af-469b-83cc-4ab8859e0e2f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.000435] env[62385]: DEBUG nova.scheduler.client.report [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 713.013777] env[62385]: DEBUG oslo_vmware.api [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5289567e-a285-3fbc-8f03-30134cb84e47, 'name': SearchDatastore_Task, 'duration_secs': 0.009216} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.014241] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.014360] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 713.018020] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.018020] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.018020] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 713.018020] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ae1fd1ef-9b1d-4227-a982-b42c7b53fc09 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.027143] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 713.027143] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 713.027827] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-021667c9-2fbb-4146-82f0-d45e95d844d3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.034012] env[62385]: DEBUG oslo_vmware.api [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Waiting for the task: (returnval){ [ 713.034012] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ae7415-640c-5ada-3fdb-73f332d4755d" [ 713.034012] env[62385]: _type = "Task" [ 713.034012] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.042120] env[62385]: DEBUG oslo_vmware.api [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ae7415-640c-5ada-3fdb-73f332d4755d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.067109] env[62385]: DEBUG nova.network.neutron [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Updating instance_info_cache with network_info: [{"id": "60df554a-a3e2-41eb-b893-16459c19bc7c", "address": "fa:16:3e:98:8e:d9", "network": {"id": "1d0e9098-63d3-4df5-a80a-06f9a5d90db6", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-303269893-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72549cb052314f149111e4e24ffb6656", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "572b7281-aad3-45fa-9cb2-fc1c70569948", "external-id": "nsx-vlan-transportzone-722", "segmentation_id": 722, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap60df554a-a3", "ovs_interfaceid": "60df554a-a3e2-41eb-b893-16459c19bc7c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.072992] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Acquiring lock "2b744879-01d2-440e-8adf-58438b922888" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 713.073219] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Lock "2b744879-01d2-440e-8adf-58438b922888" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 713.155624] env[62385]: DEBUG oslo_vmware.api [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205489, 'name': Rename_Task, 'duration_secs': 0.140068} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.155901] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 713.156164] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8e6dcf25-734d-4e46-96aa-796330cfd1ba {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.162728] env[62385]: DEBUG oslo_vmware.api [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Waiting for the task: (returnval){ [ 713.162728] env[62385]: value = "task-1205490" [ 713.162728] env[62385]: _type = "Task" [ 713.162728] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.173325] env[62385]: DEBUG oslo_vmware.api [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205490, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.196905] env[62385]: DEBUG nova.network.neutron [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.361177] env[62385]: DEBUG oslo_concurrency.lockutils [req-9b6981b4-6983-4fe0-8627-963ce878999e req-5ec560d4-f010-46b7-be06-010c88e63d75 service nova] Releasing lock "refresh_cache-4a4b254c-46a3-4db6-9c0f-f476ae8f2b69" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.508129] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.025s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 713.508745] env[62385]: ERROR nova.compute.manager [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6091e7b7-501c-4419-b688-843834c15097, please check neutron logs for more information. [ 713.508745] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Traceback (most recent call last): [ 713.508745] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 713.508745] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] self.driver.spawn(context, instance, image_meta, [ 713.508745] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 713.508745] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 713.508745] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 713.508745] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] vm_ref = self.build_virtual_machine(instance, [ 713.508745] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 713.508745] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] vif_infos = vmwarevif.get_vif_info(self._session, [ 713.508745] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 713.509107] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] for vif in network_info: [ 713.509107] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 713.509107] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] return self._sync_wrapper(fn, *args, **kwargs) [ 713.509107] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 713.509107] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] self.wait() [ 713.509107] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 713.509107] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] self[:] = self._gt.wait() [ 713.509107] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 713.509107] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] return self._exit_event.wait() [ 713.509107] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 713.509107] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] current.throw(*self._exc) [ 713.509107] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 713.509107] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] result = function(*args, **kwargs) [ 713.509621] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 713.509621] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] return func(*args, **kwargs) [ 713.509621] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 713.509621] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] raise e [ 713.509621] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 713.509621] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] nwinfo = self.network_api.allocate_for_instance( [ 713.509621] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 713.509621] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] created_port_ids = self._update_ports_for_instance( [ 713.509621] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 713.509621] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] with excutils.save_and_reraise_exception(): [ 713.509621] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 713.509621] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] self.force_reraise() [ 713.509621] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 713.510143] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] raise self.value [ 713.510143] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 713.510143] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] updated_port = self._update_port( [ 713.510143] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 713.510143] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] _ensure_no_port_binding_failure(port) [ 713.510143] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 713.510143] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] raise exception.PortBindingFailed(port_id=port['id']) [ 713.510143] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] nova.exception.PortBindingFailed: Binding failed for port 6091e7b7-501c-4419-b688-843834c15097, please check neutron logs for more information. [ 713.510143] env[62385]: ERROR nova.compute.manager [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] [ 713.510143] env[62385]: DEBUG nova.compute.utils [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Binding failed for port 6091e7b7-501c-4419-b688-843834c15097, please check neutron logs for more information. {{(pid=62385) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 713.510721] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.606s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 713.512052] env[62385]: INFO nova.compute.claims [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 713.515208] env[62385]: DEBUG nova.compute.manager [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Build of instance d3224596-4e38-4c9c-abd9-59a34bbf59c7 was re-scheduled: Binding failed for port 6091e7b7-501c-4419-b688-843834c15097, please check neutron logs for more information. {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 713.515673] env[62385]: DEBUG nova.compute.manager [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Unplugging VIFs for instance {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 713.515904] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Acquiring lock "refresh_cache-d3224596-4e38-4c9c-abd9-59a34bbf59c7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.516082] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Acquired lock "refresh_cache-d3224596-4e38-4c9c-abd9-59a34bbf59c7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.516246] env[62385]: DEBUG nova.network.neutron [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 713.545863] env[62385]: DEBUG oslo_vmware.api [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ae7415-640c-5ada-3fdb-73f332d4755d, 'name': SearchDatastore_Task, 'duration_secs': 0.009367} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.546120] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49b8a1ea-d73f-4a44-ad71-ffb9dabd1018 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.551370] env[62385]: DEBUG oslo_vmware.api [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Waiting for the task: (returnval){ [ 713.551370] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]520c487c-892d-79ef-04ff-a8f24a4de5d2" [ 713.551370] env[62385]: _type = "Task" [ 713.551370] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.559188] env[62385]: DEBUG oslo_vmware.api [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]520c487c-892d-79ef-04ff-a8f24a4de5d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.572674] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Releasing lock "refresh_cache-3e184125-28af-469b-83cc-4ab8859e0e2f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.572674] env[62385]: DEBUG nova.compute.manager [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Instance network_info: |[{"id": "60df554a-a3e2-41eb-b893-16459c19bc7c", "address": "fa:16:3e:98:8e:d9", "network": {"id": "1d0e9098-63d3-4df5-a80a-06f9a5d90db6", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-303269893-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72549cb052314f149111e4e24ffb6656", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "572b7281-aad3-45fa-9cb2-fc1c70569948", "external-id": "nsx-vlan-transportzone-722", "segmentation_id": 722, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap60df554a-a3", "ovs_interfaceid": "60df554a-a3e2-41eb-b893-16459c19bc7c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 713.572903] env[62385]: DEBUG oslo_concurrency.lockutils [req-207d236e-b8c9-43d9-bfa4-d3d0f83ec11e req-2e6ee74b-6615-4153-b3fd-c1fdc598b530 service nova] Acquired lock "refresh_cache-3e184125-28af-469b-83cc-4ab8859e0e2f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.572903] env[62385]: DEBUG nova.network.neutron [req-207d236e-b8c9-43d9-bfa4-d3d0f83ec11e req-2e6ee74b-6615-4153-b3fd-c1fdc598b530 service nova] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Refreshing network info cache for port 60df554a-a3e2-41eb-b893-16459c19bc7c {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 713.573783] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:98:8e:d9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '572b7281-aad3-45fa-9cb2-fc1c70569948', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '60df554a-a3e2-41eb-b893-16459c19bc7c', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 713.582162] env[62385]: DEBUG oslo.service.loopingcall [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 713.583209] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 713.583448] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-91495aef-6e7e-4dae-ba35-f354882a539e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.603697] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 713.603697] env[62385]: value = "task-1205491" [ 713.603697] env[62385]: _type = "Task" [ 713.603697] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.612523] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205491, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.675048] env[62385]: DEBUG oslo_vmware.api [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205490, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.698878] env[62385]: INFO nova.compute.manager [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] [instance: 1b459bac-e27d-4af5-bfc2-49898f5ced25] Took 1.02 seconds to deallocate network for instance. [ 714.038899] env[62385]: DEBUG nova.network.neutron [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 714.064085] env[62385]: DEBUG oslo_vmware.api [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]520c487c-892d-79ef-04ff-a8f24a4de5d2, 'name': SearchDatastore_Task, 'duration_secs': 0.009241} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.065114] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.065114] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69/4a4b254c-46a3-4db6-9c0f-f476ae8f2b69.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 714.065114] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7ff56bd7-2798-404f-99ef-702d00364116 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.071721] env[62385]: DEBUG oslo_vmware.api [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Waiting for the task: (returnval){ [ 714.071721] env[62385]: value = "task-1205492" [ 714.071721] env[62385]: _type = "Task" [ 714.071721] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.080848] env[62385]: DEBUG oslo_vmware.api [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205492, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.105412] env[62385]: DEBUG nova.network.neutron [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.116641] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205491, 'name': CreateVM_Task} progress is 25%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.173786] env[62385]: DEBUG oslo_vmware.api [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205490, 'name': PowerOnVM_Task, 'duration_secs': 0.539947} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.174116] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 714.174116] env[62385]: INFO nova.compute.manager [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Took 7.68 seconds to spawn the instance on the hypervisor. [ 714.174260] env[62385]: DEBUG nova.compute.manager [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 714.175089] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22f140bf-434d-4f2b-a16c-537ef4b2dbf0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.355559] env[62385]: DEBUG nova.network.neutron [req-207d236e-b8c9-43d9-bfa4-d3d0f83ec11e req-2e6ee74b-6615-4153-b3fd-c1fdc598b530 service nova] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Updated VIF entry in instance network info cache for port 60df554a-a3e2-41eb-b893-16459c19bc7c. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 714.355965] env[62385]: DEBUG nova.network.neutron [req-207d236e-b8c9-43d9-bfa4-d3d0f83ec11e req-2e6ee74b-6615-4153-b3fd-c1fdc598b530 service nova] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Updating instance_info_cache with network_info: [{"id": "60df554a-a3e2-41eb-b893-16459c19bc7c", "address": "fa:16:3e:98:8e:d9", "network": {"id": "1d0e9098-63d3-4df5-a80a-06f9a5d90db6", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-303269893-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72549cb052314f149111e4e24ffb6656", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "572b7281-aad3-45fa-9cb2-fc1c70569948", "external-id": "nsx-vlan-transportzone-722", "segmentation_id": 722, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap60df554a-a3", "ovs_interfaceid": "60df554a-a3e2-41eb-b893-16459c19bc7c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.582071] env[62385]: DEBUG oslo_vmware.api [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205492, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.475135} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.584570] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69/4a4b254c-46a3-4db6-9c0f-f476ae8f2b69.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 714.584795] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 714.585223] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ef036d15-851e-4ff7-8b3b-346077786150 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.591935] env[62385]: DEBUG oslo_vmware.api [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Waiting for the task: (returnval){ [ 714.591935] env[62385]: value = "task-1205493" [ 714.591935] env[62385]: _type = "Task" [ 714.591935] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.602279] env[62385]: DEBUG oslo_vmware.api [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205493, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.611730] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Releasing lock "refresh_cache-d3224596-4e38-4c9c-abd9-59a34bbf59c7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.611983] env[62385]: DEBUG nova.compute.manager [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 714.612129] env[62385]: DEBUG nova.compute.manager [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 714.612295] env[62385]: DEBUG nova.network.neutron [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 714.619781] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205491, 'name': CreateVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.628113] env[62385]: DEBUG nova.network.neutron [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 714.691503] env[62385]: INFO nova.compute.manager [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Took 29.66 seconds to build instance. [ 714.728271] env[62385]: INFO nova.scheduler.client.report [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] Deleted allocations for instance 1b459bac-e27d-4af5-bfc2-49898f5ced25 [ 714.859246] env[62385]: DEBUG oslo_concurrency.lockutils [req-207d236e-b8c9-43d9-bfa4-d3d0f83ec11e req-2e6ee74b-6615-4153-b3fd-c1fdc598b530 service nova] Releasing lock "refresh_cache-3e184125-28af-469b-83cc-4ab8859e0e2f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.865089] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dca5e925-de3e-45f9-b96e-ad0241811c60 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.872608] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f9c2bff-5ba4-4947-af32-64933cc88cbd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.902209] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25319367-5d2b-4978-9365-f68b7c8adb0a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.909933] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc0f7422-9eba-4cfc-a137-a7867a69ae4e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.922952] env[62385]: DEBUG nova.compute.provider_tree [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 715.102797] env[62385]: DEBUG oslo_vmware.api [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205493, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062237} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.103071] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 715.103816] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f315edb1-2e51-453d-be0f-f0d2e82db1c4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.125301] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Reconfiguring VM instance instance-0000002a to attach disk [datastore1] 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69/4a4b254c-46a3-4db6-9c0f-f476ae8f2b69.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 715.128202] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6eadb6ce-b3a6-41cd-af80-ce9ab2eae173 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.144732] env[62385]: DEBUG nova.network.neutron [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.151540] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205491, 'name': CreateVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.153092] env[62385]: DEBUG oslo_vmware.api [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Waiting for the task: (returnval){ [ 715.153092] env[62385]: value = "task-1205494" [ 715.153092] env[62385]: _type = "Task" [ 715.153092] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.163283] env[62385]: DEBUG oslo_vmware.api [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205494, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.196013] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ad7659cf-1bcd-47ef-8d07-f2ce745b4fd7 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Lock "1ddb4d6d-3872-49db-bb40-e21721241e89" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 145.310s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.236015] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51d02e5b-87d1-4626-9611-2c573cf5fda8 tempest-InstanceActionsTestJSON-791817975 tempest-InstanceActionsTestJSON-791817975-project-member] Lock "1b459bac-e27d-4af5-bfc2-49898f5ced25" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 151.908s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.427930] env[62385]: DEBUG nova.scheduler.client.report [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 715.626256] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205491, 'name': CreateVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.648158] env[62385]: INFO nova.compute.manager [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] [instance: d3224596-4e38-4c9c-abd9-59a34bbf59c7] Took 1.03 seconds to deallocate network for instance. [ 715.666732] env[62385]: DEBUG oslo_vmware.api [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205494, 'name': ReconfigVM_Task, 'duration_secs': 0.39708} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.667587] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Reconfigured VM instance instance-0000002a to attach disk [datastore1] 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69/4a4b254c-46a3-4db6-9c0f-f476ae8f2b69.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 715.669326] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a3ebb045-5fa1-41e3-8aaa-18330cfa904a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.677280] env[62385]: DEBUG oslo_vmware.api [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Waiting for the task: (returnval){ [ 715.677280] env[62385]: value = "task-1205495" [ 715.677280] env[62385]: _type = "Task" [ 715.677280] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.683329] env[62385]: DEBUG oslo_vmware.api [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205495, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.699248] env[62385]: DEBUG nova.compute.manager [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 715.738934] env[62385]: DEBUG nova.compute.manager [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 715.935052] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.424s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.935742] env[62385]: DEBUG nova.compute.manager [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 715.939133] env[62385]: DEBUG oslo_concurrency.lockutils [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.271s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.127030] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205491, 'name': CreateVM_Task, 'duration_secs': 2.210728} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.127258] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 716.128054] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.128225] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.128594] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 716.128872] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2f1962e-8bb6-404c-8fb2-9a63d8f91e68 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.133555] env[62385]: DEBUG oslo_vmware.api [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Waiting for the task: (returnval){ [ 716.133555] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]520a3c08-a1f0-141d-26cd-58e547a5412f" [ 716.133555] env[62385]: _type = "Task" [ 716.133555] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.141716] env[62385]: DEBUG oslo_vmware.api [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]520a3c08-a1f0-141d-26cd-58e547a5412f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.186443] env[62385]: DEBUG oslo_vmware.api [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205495, 'name': Rename_Task, 'duration_secs': 0.140621} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.186743] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 716.186991] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8af2d1c2-baf4-481f-9df3-2387b0f42d28 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.193990] env[62385]: DEBUG oslo_vmware.api [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Waiting for the task: (returnval){ [ 716.193990] env[62385]: value = "task-1205496" [ 716.193990] env[62385]: _type = "Task" [ 716.193990] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.201921] env[62385]: DEBUG oslo_vmware.api [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205496, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.229174] env[62385]: DEBUG oslo_concurrency.lockutils [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.263137] env[62385]: DEBUG oslo_concurrency.lockutils [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.448023] env[62385]: DEBUG nova.compute.utils [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 716.450637] env[62385]: DEBUG nova.compute.manager [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 716.451817] env[62385]: DEBUG nova.network.neutron [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 716.500685] env[62385]: DEBUG nova.policy [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '825ffbadf064473aa5eda331914306f2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7ac0c8638f064daba042ded4de8c1cc5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 716.646882] env[62385]: DEBUG oslo_vmware.api [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]520a3c08-a1f0-141d-26cd-58e547a5412f, 'name': SearchDatastore_Task, 'duration_secs': 0.009525} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.647201] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.647438] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 716.647705] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.647853] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.648156] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 716.648456] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b9385614-558f-4083-94de-badfd8b1797d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.661303] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 716.661402] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 716.662288] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2244c2f3-2467-49b7-a0f6-28f27fd6a245 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.667797] env[62385]: DEBUG oslo_vmware.api [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Waiting for the task: (returnval){ [ 716.667797] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52dc8d49-36bc-2a76-9fba-130ec6dc3b72" [ 716.667797] env[62385]: _type = "Task" [ 716.667797] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.678306] env[62385]: DEBUG oslo_vmware.api [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52dc8d49-36bc-2a76-9fba-130ec6dc3b72, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.692033] env[62385]: INFO nova.scheduler.client.report [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Deleted allocations for instance d3224596-4e38-4c9c-abd9-59a34bbf59c7 [ 716.707321] env[62385]: DEBUG oslo_vmware.api [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205496, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.872605] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e6dd3ca-5cc5-4c33-bb3b-ad40d6f7753a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.885999] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d5607e6-cdfc-4ce9-a587-2d39e4c51d38 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.918514] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe94c489-6f49-4305-a8c0-3a2393f8c799 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.926569] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05ac2561-3a67-4597-8ece-4d4a779f1118 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.942954] env[62385]: DEBUG nova.compute.provider_tree [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 716.951172] env[62385]: DEBUG nova.compute.manager [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 717.091588] env[62385]: DEBUG nova.network.neutron [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Successfully created port: 72d846e7-a597-4eb7-87ea-0d4309ceb42b {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 717.180023] env[62385]: DEBUG oslo_vmware.api [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52dc8d49-36bc-2a76-9fba-130ec6dc3b72, 'name': SearchDatastore_Task, 'duration_secs': 0.016108} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.180885] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a011f280-2595-4b60-96ad-72212147260e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.186347] env[62385]: DEBUG oslo_vmware.api [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Waiting for the task: (returnval){ [ 717.186347] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52f33949-7b38-a778-d0b7-f4f1528a93ce" [ 717.186347] env[62385]: _type = "Task" [ 717.186347] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.194377] env[62385]: DEBUG oslo_vmware.api [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52f33949-7b38-a778-d0b7-f4f1528a93ce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.204327] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e8fbdd6f-6e03-466f-b367-ee4ba28bb48a tempest-MigrationsAdminTest-91861838 tempest-MigrationsAdminTest-91861838-project-member] Lock "d3224596-4e38-4c9c-abd9-59a34bbf59c7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 153.794s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.210101] env[62385]: DEBUG oslo_vmware.api [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205496, 'name': PowerOnVM_Task, 'duration_secs': 0.708061} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.210350] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 717.210541] env[62385]: INFO nova.compute.manager [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Took 8.29 seconds to spawn the instance on the hypervisor. [ 717.210712] env[62385]: DEBUG nova.compute.manager [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 717.211558] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e175acff-951d-4913-8b2e-6bf1d05331f3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.446294] env[62385]: DEBUG nova.scheduler.client.report [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 717.697914] env[62385]: DEBUG oslo_vmware.api [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52f33949-7b38-a778-d0b7-f4f1528a93ce, 'name': SearchDatastore_Task, 'duration_secs': 0.031194} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.698202] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.698453] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 3e184125-28af-469b-83cc-4ab8859e0e2f/3e184125-28af-469b-83cc-4ab8859e0e2f.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 717.698712] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0ef097b4-2335-430c-b67c-c5fea4727434 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.705585] env[62385]: DEBUG oslo_vmware.api [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Waiting for the task: (returnval){ [ 717.705585] env[62385]: value = "task-1205497" [ 717.705585] env[62385]: _type = "Task" [ 717.705585] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.708833] env[62385]: DEBUG nova.compute.manager [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 717.716108] env[62385]: DEBUG oslo_vmware.api [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205497, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.733972] env[62385]: INFO nova.compute.manager [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Took 30.73 seconds to build instance. [ 717.952093] env[62385]: DEBUG oslo_concurrency.lockutils [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.013s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.952757] env[62385]: ERROR nova.compute.manager [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ad28927c-3e0a-47c2-b304-43949d2afc0d, please check neutron logs for more information. [ 717.952757] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Traceback (most recent call last): [ 717.952757] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 717.952757] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] self.driver.spawn(context, instance, image_meta, [ 717.952757] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 717.952757] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] self._vmops.spawn(context, instance, image_meta, injected_files, [ 717.952757] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 717.952757] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] vm_ref = self.build_virtual_machine(instance, [ 717.952757] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 717.952757] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] vif_infos = vmwarevif.get_vif_info(self._session, [ 717.952757] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 717.953173] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] for vif in network_info: [ 717.953173] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 717.953173] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] return self._sync_wrapper(fn, *args, **kwargs) [ 717.953173] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 717.953173] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] self.wait() [ 717.953173] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 717.953173] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] self[:] = self._gt.wait() [ 717.953173] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 717.953173] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] return self._exit_event.wait() [ 717.953173] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 717.953173] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] current.throw(*self._exc) [ 717.953173] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 717.953173] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] result = function(*args, **kwargs) [ 717.953575] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 717.953575] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] return func(*args, **kwargs) [ 717.953575] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 717.953575] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] raise e [ 717.953575] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 717.953575] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] nwinfo = self.network_api.allocate_for_instance( [ 717.953575] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 717.953575] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] created_port_ids = self._update_ports_for_instance( [ 717.953575] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 717.953575] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] with excutils.save_and_reraise_exception(): [ 717.953575] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 717.953575] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] self.force_reraise() [ 717.953575] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 717.954029] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] raise self.value [ 717.954029] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 717.954029] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] updated_port = self._update_port( [ 717.954029] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 717.954029] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] _ensure_no_port_binding_failure(port) [ 717.954029] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 717.954029] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] raise exception.PortBindingFailed(port_id=port['id']) [ 717.954029] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] nova.exception.PortBindingFailed: Binding failed for port ad28927c-3e0a-47c2-b304-43949d2afc0d, please check neutron logs for more information. [ 717.954029] env[62385]: ERROR nova.compute.manager [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] [ 717.954029] env[62385]: DEBUG nova.compute.utils [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Binding failed for port ad28927c-3e0a-47c2-b304-43949d2afc0d, please check neutron logs for more information. {{(pid=62385) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 717.955321] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.653s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.956926] env[62385]: INFO nova.compute.claims [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 717.959645] env[62385]: DEBUG nova.compute.manager [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Build of instance 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261 was re-scheduled: Binding failed for port ad28927c-3e0a-47c2-b304-43949d2afc0d, please check neutron logs for more information. {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 717.960107] env[62385]: DEBUG nova.compute.manager [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Unplugging VIFs for instance {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 717.960396] env[62385]: DEBUG oslo_concurrency.lockutils [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] Acquiring lock "refresh_cache-4c96b9c6-40b9-4dea-8dd3-5e8e6372c261" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.960476] env[62385]: DEBUG oslo_concurrency.lockutils [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] Acquired lock "refresh_cache-4c96b9c6-40b9-4dea-8dd3-5e8e6372c261" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.960625] env[62385]: DEBUG nova.network.neutron [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 717.962152] env[62385]: DEBUG nova.compute.manager [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 717.985201] env[62385]: DEBUG nova.virt.hardware [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 717.985442] env[62385]: DEBUG nova.virt.hardware [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 717.985695] env[62385]: DEBUG nova.virt.hardware [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 717.985885] env[62385]: DEBUG nova.virt.hardware [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 717.986030] env[62385]: DEBUG nova.virt.hardware [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 717.986189] env[62385]: DEBUG nova.virt.hardware [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 717.986499] env[62385]: DEBUG nova.virt.hardware [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 717.986575] env[62385]: DEBUG nova.virt.hardware [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 717.986766] env[62385]: DEBUG nova.virt.hardware [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 717.986886] env[62385]: DEBUG nova.virt.hardware [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 717.987099] env[62385]: DEBUG nova.virt.hardware [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 717.988216] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5a66b15-91c7-4989-97c7-7a091f3eeb4f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.996536] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-143c2466-939b-4ee9-9798-abca6e123144 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.218884] env[62385]: DEBUG oslo_vmware.api [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205497, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.233869] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 718.235598] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d3ecec9b-e5f5-461b-b0a8-81f7a426d638 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Lock "4a4b254c-46a3-4db6-9c0f-f476ae8f2b69" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 146.685s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.494343] env[62385]: DEBUG nova.network.neutron [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 718.621909] env[62385]: DEBUG nova.network.neutron [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.721713] env[62385]: DEBUG oslo_vmware.api [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205497, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.677564} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.722058] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 3e184125-28af-469b-83cc-4ab8859e0e2f/3e184125-28af-469b-83cc-4ab8859e0e2f.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 718.722278] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 718.722604] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4fcebe30-223c-4567-885a-3205401dc81f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.731221] env[62385]: DEBUG oslo_vmware.api [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Waiting for the task: (returnval){ [ 718.731221] env[62385]: value = "task-1205498" [ 718.731221] env[62385]: _type = "Task" [ 718.731221] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.743490] env[62385]: DEBUG nova.compute.manager [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 718.748612] env[62385]: DEBUG oslo_vmware.api [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205498, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.822632] env[62385]: DEBUG nova.compute.manager [req-7488f0f9-3616-4bc7-9ada-e2bcff51582a req-4c5e6032-0d3e-4733-9d39-39da49f13f0d service nova] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Received event network-vif-plugged-72d846e7-a597-4eb7-87ea-0d4309ceb42b {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 718.822882] env[62385]: DEBUG oslo_concurrency.lockutils [req-7488f0f9-3616-4bc7-9ada-e2bcff51582a req-4c5e6032-0d3e-4733-9d39-39da49f13f0d service nova] Acquiring lock "33387505-c576-488b-8c9c-b064fe81a7d7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 718.823161] env[62385]: DEBUG oslo_concurrency.lockutils [req-7488f0f9-3616-4bc7-9ada-e2bcff51582a req-4c5e6032-0d3e-4733-9d39-39da49f13f0d service nova] Lock "33387505-c576-488b-8c9c-b064fe81a7d7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.823981] env[62385]: DEBUG oslo_concurrency.lockutils [req-7488f0f9-3616-4bc7-9ada-e2bcff51582a req-4c5e6032-0d3e-4733-9d39-39da49f13f0d service nova] Lock "33387505-c576-488b-8c9c-b064fe81a7d7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.823981] env[62385]: DEBUG nova.compute.manager [req-7488f0f9-3616-4bc7-9ada-e2bcff51582a req-4c5e6032-0d3e-4733-9d39-39da49f13f0d service nova] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] No waiting events found dispatching network-vif-plugged-72d846e7-a597-4eb7-87ea-0d4309ceb42b {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 718.823981] env[62385]: WARNING nova.compute.manager [req-7488f0f9-3616-4bc7-9ada-e2bcff51582a req-4c5e6032-0d3e-4733-9d39-39da49f13f0d service nova] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Received unexpected event network-vif-plugged-72d846e7-a597-4eb7-87ea-0d4309ceb42b for instance with vm_state building and task_state spawning. [ 718.963978] env[62385]: DEBUG nova.network.neutron [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Successfully updated port: 72d846e7-a597-4eb7-87ea-0d4309ceb42b {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 719.123523] env[62385]: DEBUG oslo_concurrency.lockutils [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] Releasing lock "refresh_cache-4c96b9c6-40b9-4dea-8dd3-5e8e6372c261" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.124229] env[62385]: DEBUG nova.compute.manager [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 719.124229] env[62385]: DEBUG nova.compute.manager [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 719.124229] env[62385]: DEBUG nova.network.neutron [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 719.141065] env[62385]: DEBUG nova.network.neutron [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 719.241478] env[62385]: DEBUG oslo_vmware.api [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205498, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069126} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.243986] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 719.245025] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f07102c8-edf8-4a0c-8473-095d13702570 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.269562] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Reconfiguring VM instance instance-0000002b to attach disk [datastore1] 3e184125-28af-469b-83cc-4ab8859e0e2f/3e184125-28af-469b-83cc-4ab8859e0e2f.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 719.274112] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9069a892-9336-4966-9244-798af0bde041 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.293967] env[62385]: DEBUG oslo_vmware.api [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Waiting for the task: (returnval){ [ 719.293967] env[62385]: value = "task-1205499" [ 719.293967] env[62385]: _type = "Task" [ 719.293967] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.303440] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.306623] env[62385]: DEBUG oslo_vmware.api [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205499, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.348275] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bd539ac-bacf-484d-bb64-0730b780202d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.359486] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a5a9ece-43c2-4de6-bece-61d6e30c5995 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.401117] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acb0a961-34e1-49f2-b189-aca17105b838 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.409391] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93461c08-bbd8-4c1d-b124-50524df0e307 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.423764] env[62385]: DEBUG nova.compute.provider_tree [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 719.466171] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquiring lock "refresh_cache-33387505-c576-488b-8c9c-b064fe81a7d7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.466319] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquired lock "refresh_cache-33387505-c576-488b-8c9c-b064fe81a7d7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.466538] env[62385]: DEBUG nova.network.neutron [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 719.647524] env[62385]: DEBUG nova.network.neutron [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.804293] env[62385]: DEBUG oslo_vmware.api [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205499, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.927124] env[62385]: DEBUG nova.scheduler.client.report [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 720.019697] env[62385]: DEBUG nova.network.neutron [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 720.154021] env[62385]: INFO nova.compute.manager [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] [instance: 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261] Took 1.03 seconds to deallocate network for instance. [ 720.180891] env[62385]: DEBUG nova.network.neutron [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Updating instance_info_cache with network_info: [{"id": "72d846e7-a597-4eb7-87ea-0d4309ceb42b", "address": "fa:16:3e:74:68:57", "network": {"id": "b0a3dea4-1925-4920-8a8e-9dc66733d1da", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1951021425-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ac0c8638f064daba042ded4de8c1cc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee20e439-fed9-490e-97dd-f3c886977ae1", "external-id": "nsx-vlan-transportzone-357", "segmentation_id": 357, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72d846e7-a5", "ovs_interfaceid": "72d846e7-a597-4eb7-87ea-0d4309ceb42b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.305312] env[62385]: DEBUG oslo_vmware.api [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205499, 'name': ReconfigVM_Task, 'duration_secs': 0.537194} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.305573] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Reconfigured VM instance instance-0000002b to attach disk [datastore1] 3e184125-28af-469b-83cc-4ab8859e0e2f/3e184125-28af-469b-83cc-4ab8859e0e2f.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 720.306626] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-13671686-f969-4b00-ba63-a8e838f2d090 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.312388] env[62385]: DEBUG oslo_vmware.api [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Waiting for the task: (returnval){ [ 720.312388] env[62385]: value = "task-1205500" [ 720.312388] env[62385]: _type = "Task" [ 720.312388] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.319792] env[62385]: DEBUG oslo_vmware.api [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205500, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.433515] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.478s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.434065] env[62385]: DEBUG nova.compute.manager [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 720.436809] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.955s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 720.685484] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Releasing lock "refresh_cache-33387505-c576-488b-8c9c-b064fe81a7d7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.687220] env[62385]: DEBUG nova.compute.manager [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Instance network_info: |[{"id": "72d846e7-a597-4eb7-87ea-0d4309ceb42b", "address": "fa:16:3e:74:68:57", "network": {"id": "b0a3dea4-1925-4920-8a8e-9dc66733d1da", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1951021425-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ac0c8638f064daba042ded4de8c1cc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee20e439-fed9-490e-97dd-f3c886977ae1", "external-id": "nsx-vlan-transportzone-357", "segmentation_id": 357, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72d846e7-a5", "ovs_interfaceid": "72d846e7-a597-4eb7-87ea-0d4309ceb42b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 720.687370] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:74:68:57', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ee20e439-fed9-490e-97dd-f3c886977ae1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '72d846e7-a597-4eb7-87ea-0d4309ceb42b', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 720.693685] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Creating folder: Project (7ac0c8638f064daba042ded4de8c1cc5). Parent ref: group-v261107. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 720.694615] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-06c303b2-871f-4a7f-a5b1-95f998fa14c3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.705503] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Created folder: Project (7ac0c8638f064daba042ded4de8c1cc5) in parent group-v261107. [ 720.705761] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Creating folder: Instances. Parent ref: group-v261125. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 720.705915] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a094fb37-6409-4296-9184-c32151cdb731 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.715697] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Created folder: Instances in parent group-v261125. [ 720.717278] env[62385]: DEBUG oslo.service.loopingcall [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 720.717278] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 720.717278] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e0c442e0-b77c-4ca0-a378-f17bfd17fd73 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.735126] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 720.735126] env[62385]: value = "task-1205503" [ 720.735126] env[62385]: _type = "Task" [ 720.735126] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.744642] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205503, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.825142] env[62385]: DEBUG oslo_vmware.api [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205500, 'name': Rename_Task, 'duration_secs': 0.154293} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.825142] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 720.825142] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-598b354c-041b-476a-befa-f32ac4500092 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.829433] env[62385]: DEBUG oslo_vmware.api [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Waiting for the task: (returnval){ [ 720.829433] env[62385]: value = "task-1205504" [ 720.829433] env[62385]: _type = "Task" [ 720.829433] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.836826] env[62385]: DEBUG oslo_vmware.api [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205504, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.847282] env[62385]: DEBUG nova.compute.manager [req-46948348-5a88-453f-a266-30eb6715f85b req-e01a8e2d-197f-447f-a193-733de74e10f8 service nova] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Received event network-changed-72d846e7-a597-4eb7-87ea-0d4309ceb42b {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 720.847943] env[62385]: DEBUG nova.compute.manager [req-46948348-5a88-453f-a266-30eb6715f85b req-e01a8e2d-197f-447f-a193-733de74e10f8 service nova] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Refreshing instance network info cache due to event network-changed-72d846e7-a597-4eb7-87ea-0d4309ceb42b. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 720.847943] env[62385]: DEBUG oslo_concurrency.lockutils [req-46948348-5a88-453f-a266-30eb6715f85b req-e01a8e2d-197f-447f-a193-733de74e10f8 service nova] Acquiring lock "refresh_cache-33387505-c576-488b-8c9c-b064fe81a7d7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.847943] env[62385]: DEBUG oslo_concurrency.lockutils [req-46948348-5a88-453f-a266-30eb6715f85b req-e01a8e2d-197f-447f-a193-733de74e10f8 service nova] Acquired lock "refresh_cache-33387505-c576-488b-8c9c-b064fe81a7d7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.848262] env[62385]: DEBUG nova.network.neutron [req-46948348-5a88-453f-a266-30eb6715f85b req-e01a8e2d-197f-447f-a193-733de74e10f8 service nova] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Refreshing network info cache for port 72d846e7-a597-4eb7-87ea-0d4309ceb42b {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 720.944482] env[62385]: DEBUG nova.compute.utils [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 720.946529] env[62385]: DEBUG nova.compute.manager [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 720.946529] env[62385]: DEBUG nova.network.neutron [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 720.989211] env[62385]: DEBUG nova.policy [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '825ffbadf064473aa5eda331914306f2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7ac0c8638f064daba042ded4de8c1cc5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 721.199654] env[62385]: INFO nova.scheduler.client.report [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] Deleted allocations for instance 4c96b9c6-40b9-4dea-8dd3-5e8e6372c261 [ 721.248481] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205503, 'name': CreateVM_Task, 'duration_secs': 0.348105} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.252029] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 721.252029] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 721.252029] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.252029] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 721.252029] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-262df8f0-9712-4b8e-91d8-03b6fe9a5844 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.262846] env[62385]: DEBUG oslo_vmware.api [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 721.262846] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52641436-674b-2c46-db50-b670b9449cf9" [ 721.262846] env[62385]: _type = "Task" [ 721.262846] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.272785] env[62385]: DEBUG oslo_vmware.api [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52641436-674b-2c46-db50-b670b9449cf9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.316927] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9673cdd1-2f55-468e-b61c-b36f51523271 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.326104] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9698ade7-e877-40e8-8373-9e6e6253f3ff {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.339215] env[62385]: DEBUG oslo_vmware.api [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205504, 'name': PowerOnVM_Task, 'duration_secs': 0.452996} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.362921] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 721.363130] env[62385]: INFO nova.compute.manager [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Took 9.91 seconds to spawn the instance on the hypervisor. [ 721.363298] env[62385]: DEBUG nova.compute.manager [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 721.366354] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-099e5737-890a-4da8-9175-c70ab2ba9614 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.369491] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42eeea60-2750-4491-ab0f-279bed2eac49 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.379253] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cba9df76-3928-4bd8-a078-f2af48aad726 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.395907] env[62385]: DEBUG nova.compute.provider_tree [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 721.397810] env[62385]: DEBUG nova.network.neutron [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Successfully created port: 65f0aa57-f11b-49fb-8889-1fab7629b3d0 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 721.451323] env[62385]: DEBUG nova.compute.manager [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 721.616271] env[62385]: DEBUG nova.network.neutron [req-46948348-5a88-453f-a266-30eb6715f85b req-e01a8e2d-197f-447f-a193-733de74e10f8 service nova] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Updated VIF entry in instance network info cache for port 72d846e7-a597-4eb7-87ea-0d4309ceb42b. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 721.616271] env[62385]: DEBUG nova.network.neutron [req-46948348-5a88-453f-a266-30eb6715f85b req-e01a8e2d-197f-447f-a193-733de74e10f8 service nova] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Updating instance_info_cache with network_info: [{"id": "72d846e7-a597-4eb7-87ea-0d4309ceb42b", "address": "fa:16:3e:74:68:57", "network": {"id": "b0a3dea4-1925-4920-8a8e-9dc66733d1da", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1951021425-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ac0c8638f064daba042ded4de8c1cc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee20e439-fed9-490e-97dd-f3c886977ae1", "external-id": "nsx-vlan-transportzone-357", "segmentation_id": 357, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72d846e7-a5", "ovs_interfaceid": "72d846e7-a597-4eb7-87ea-0d4309ceb42b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.712729] env[62385]: DEBUG oslo_concurrency.lockutils [None req-55ee4496-2589-49b7-b3f0-3c6ce7eaa022 tempest-ServersTestFqdnHostnames-1192766185 tempest-ServersTestFqdnHostnames-1192766185-project-member] Lock "4c96b9c6-40b9-4dea-8dd3-5e8e6372c261" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 157.565s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.774617] env[62385]: DEBUG oslo_vmware.api [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52641436-674b-2c46-db50-b670b9449cf9, 'name': SearchDatastore_Task, 'duration_secs': 0.017139} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.774617] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.774914] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 721.775166] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 721.775314] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.775522] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 721.775819] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-25c9565a-9313-4a2a-8925-f5ec4a1f480d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.786984] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 721.787198] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 721.787939] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-855a52b2-7a6d-45a0-963b-faa68924de44 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.793644] env[62385]: DEBUG oslo_vmware.api [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 721.793644] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52d6017c-3401-59e3-f5c7-fceed01f0166" [ 721.793644] env[62385]: _type = "Task" [ 721.793644] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.801286] env[62385]: DEBUG oslo_vmware.api [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52d6017c-3401-59e3-f5c7-fceed01f0166, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.903062] env[62385]: DEBUG nova.scheduler.client.report [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 721.907963] env[62385]: INFO nova.compute.manager [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Took 32.87 seconds to build instance. [ 722.121039] env[62385]: DEBUG oslo_concurrency.lockutils [req-46948348-5a88-453f-a266-30eb6715f85b req-e01a8e2d-197f-447f-a193-733de74e10f8 service nova] Releasing lock "refresh_cache-33387505-c576-488b-8c9c-b064fe81a7d7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.215327] env[62385]: DEBUG nova.compute.manager [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 722.304458] env[62385]: DEBUG oslo_vmware.api [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52d6017c-3401-59e3-f5c7-fceed01f0166, 'name': SearchDatastore_Task, 'duration_secs': 0.029399} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.305286] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0a75b31-ec3d-4973-9fd3-eba10816e040 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.310871] env[62385]: DEBUG oslo_vmware.api [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 722.310871] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e25f5c-a9f5-6b03-d505-b357db803d3f" [ 722.310871] env[62385]: _type = "Task" [ 722.310871] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.318229] env[62385]: DEBUG oslo_vmware.api [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e25f5c-a9f5-6b03-d505-b357db803d3f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.409551] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.973s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.411433] env[62385]: ERROR nova.compute.manager [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7e3b5847-c00b-4e68-b344-f4e90204abf6, please check neutron logs for more information. [ 722.411433] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Traceback (most recent call last): [ 722.411433] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 722.411433] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] self.driver.spawn(context, instance, image_meta, [ 722.411433] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 722.411433] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 722.411433] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 722.411433] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] vm_ref = self.build_virtual_machine(instance, [ 722.411433] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 722.411433] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] vif_infos = vmwarevif.get_vif_info(self._session, [ 722.411433] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 722.411802] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] for vif in network_info: [ 722.411802] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 722.411802] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] return self._sync_wrapper(fn, *args, **kwargs) [ 722.411802] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 722.411802] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] self.wait() [ 722.411802] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 722.411802] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] self[:] = self._gt.wait() [ 722.411802] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 722.411802] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] return self._exit_event.wait() [ 722.411802] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 722.411802] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] current.throw(*self._exc) [ 722.411802] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 722.411802] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] result = function(*args, **kwargs) [ 722.412227] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 722.412227] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] return func(*args, **kwargs) [ 722.412227] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 722.412227] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] raise e [ 722.412227] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 722.412227] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] nwinfo = self.network_api.allocate_for_instance( [ 722.412227] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 722.412227] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] created_port_ids = self._update_ports_for_instance( [ 722.412227] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 722.412227] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] with excutils.save_and_reraise_exception(): [ 722.412227] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 722.412227] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] self.force_reraise() [ 722.412227] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 722.412662] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] raise self.value [ 722.412662] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 722.412662] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] updated_port = self._update_port( [ 722.412662] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 722.412662] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] _ensure_no_port_binding_failure(port) [ 722.412662] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 722.412662] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] raise exception.PortBindingFailed(port_id=port['id']) [ 722.412662] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] nova.exception.PortBindingFailed: Binding failed for port 7e3b5847-c00b-4e68-b344-f4e90204abf6, please check neutron logs for more information. [ 722.412662] env[62385]: ERROR nova.compute.manager [instance: 866451cc-96c5-433a-a903-9faa1eed538a] [ 722.412662] env[62385]: DEBUG nova.compute.utils [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Binding failed for port 7e3b5847-c00b-4e68-b344-f4e90204abf6, please check neutron logs for more information. {{(pid=62385) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 722.413514] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d6d4d64d-01a5-4d3a-a700-9ba292bdf8b8 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Lock "3e184125-28af-469b-83cc-4ab8859e0e2f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 148.551s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.413736] env[62385]: DEBUG oslo_concurrency.lockutils [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.696s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 722.415253] env[62385]: INFO nova.compute.claims [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 722.417890] env[62385]: DEBUG nova.compute.manager [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Build of instance 866451cc-96c5-433a-a903-9faa1eed538a was re-scheduled: Binding failed for port 7e3b5847-c00b-4e68-b344-f4e90204abf6, please check neutron logs for more information. {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 722.418320] env[62385]: DEBUG nova.compute.manager [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Unplugging VIFs for instance {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 722.418541] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquiring lock "refresh_cache-866451cc-96c5-433a-a903-9faa1eed538a" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 722.418754] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquired lock "refresh_cache-866451cc-96c5-433a-a903-9faa1eed538a" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.418933] env[62385]: DEBUG nova.network.neutron [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 722.458622] env[62385]: DEBUG nova.compute.manager [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 722.489938] env[62385]: DEBUG nova.virt.hardware [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 722.490301] env[62385]: DEBUG nova.virt.hardware [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 722.490570] env[62385]: DEBUG nova.virt.hardware [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 722.490831] env[62385]: DEBUG nova.virt.hardware [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 722.491049] env[62385]: DEBUG nova.virt.hardware [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 722.491330] env[62385]: DEBUG nova.virt.hardware [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 722.491665] env[62385]: DEBUG nova.virt.hardware [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 722.491842] env[62385]: DEBUG nova.virt.hardware [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 722.492028] env[62385]: DEBUG nova.virt.hardware [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 722.492195] env[62385]: DEBUG nova.virt.hardware [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 722.492365] env[62385]: DEBUG nova.virt.hardware [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 722.493558] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d625e80-00b7-45e5-9f73-5b6fec56de61 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.501844] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ca27327-1ccb-4cfd-a0ea-4b659a6fdb38 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.737292] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 722.822142] env[62385]: DEBUG oslo_vmware.api [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e25f5c-a9f5-6b03-d505-b357db803d3f, 'name': SearchDatastore_Task, 'duration_secs': 0.016378} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.822591] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.823287] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 33387505-c576-488b-8c9c-b064fe81a7d7/33387505-c576-488b-8c9c-b064fe81a7d7.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 722.823503] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c0b27184-e1bc-48d6-b1a1-f4040c88dd7c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.831469] env[62385]: DEBUG oslo_vmware.api [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 722.831469] env[62385]: value = "task-1205505" [ 722.831469] env[62385]: _type = "Task" [ 722.831469] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.840665] env[62385]: DEBUG oslo_vmware.api [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205505, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.929362] env[62385]: DEBUG nova.compute.manager [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 722.956211] env[62385]: DEBUG nova.network.neutron [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 723.139460] env[62385]: DEBUG nova.network.neutron [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.341989] env[62385]: DEBUG oslo_vmware.api [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205505, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.455206} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.345234] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 33387505-c576-488b-8c9c-b064fe81a7d7/33387505-c576-488b-8c9c-b064fe81a7d7.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 723.345234] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 723.345234] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f0cf12a5-a3eb-4238-8eae-579f7d965630 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.352056] env[62385]: DEBUG oslo_vmware.api [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 723.352056] env[62385]: value = "task-1205506" [ 723.352056] env[62385]: _type = "Task" [ 723.352056] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.358542] env[62385]: DEBUG oslo_vmware.api [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205506, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.444038] env[62385]: DEBUG nova.network.neutron [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Successfully updated port: 65f0aa57-f11b-49fb-8889-1fab7629b3d0 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 723.451234] env[62385]: DEBUG nova.compute.manager [req-08a5e8dc-a4b9-4926-b214-c92f19a9c47e req-d7cc7a0b-b032-4ab9-ba3d-bf6df43bd13c service nova] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Received event network-vif-plugged-65f0aa57-f11b-49fb-8889-1fab7629b3d0 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 723.451440] env[62385]: DEBUG oslo_concurrency.lockutils [req-08a5e8dc-a4b9-4926-b214-c92f19a9c47e req-d7cc7a0b-b032-4ab9-ba3d-bf6df43bd13c service nova] Acquiring lock "f1d33401-35f5-4d79-abb4-26dc6faa784e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.451644] env[62385]: DEBUG oslo_concurrency.lockutils [req-08a5e8dc-a4b9-4926-b214-c92f19a9c47e req-d7cc7a0b-b032-4ab9-ba3d-bf6df43bd13c service nova] Lock "f1d33401-35f5-4d79-abb4-26dc6faa784e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.451802] env[62385]: DEBUG oslo_concurrency.lockutils [req-08a5e8dc-a4b9-4926-b214-c92f19a9c47e req-d7cc7a0b-b032-4ab9-ba3d-bf6df43bd13c service nova] Lock "f1d33401-35f5-4d79-abb4-26dc6faa784e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.451960] env[62385]: DEBUG nova.compute.manager [req-08a5e8dc-a4b9-4926-b214-c92f19a9c47e req-d7cc7a0b-b032-4ab9-ba3d-bf6df43bd13c service nova] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] No waiting events found dispatching network-vif-plugged-65f0aa57-f11b-49fb-8889-1fab7629b3d0 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 723.453301] env[62385]: WARNING nova.compute.manager [req-08a5e8dc-a4b9-4926-b214-c92f19a9c47e req-d7cc7a0b-b032-4ab9-ba3d-bf6df43bd13c service nova] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Received unexpected event network-vif-plugged-65f0aa57-f11b-49fb-8889-1fab7629b3d0 for instance with vm_state building and task_state spawning. [ 723.465452] env[62385]: DEBUG oslo_concurrency.lockutils [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.642807] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Releasing lock "refresh_cache-866451cc-96c5-433a-a903-9faa1eed538a" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 723.643215] env[62385]: DEBUG nova.compute.manager [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62385) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 723.643426] env[62385]: DEBUG nova.compute.manager [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 723.643598] env[62385]: DEBUG nova.network.neutron [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 723.779666] env[62385]: DEBUG nova.network.neutron [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 723.785450] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eaae27e-6f23-4cc1-b21f-e8eec74b819b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.794023] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d45e06b6-4617-41ca-882b-ae2b4e455ba6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.824905] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2084473c-9168-4973-a1b0-00bcbff3f278 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.832241] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87fb4804-10f9-4274-9796-4ccbef949822 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.847404] env[62385]: DEBUG nova.compute.provider_tree [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 723.858042] env[62385]: DEBUG oslo_vmware.api [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205506, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064083} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.858965] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 723.860180] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b722653-2aba-4268-8147-f0572b17a8d9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.886199] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Reconfiguring VM instance instance-0000002c to attach disk [datastore2] 33387505-c576-488b-8c9c-b064fe81a7d7/33387505-c576-488b-8c9c-b064fe81a7d7.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 723.886907] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b6e01ebf-27d4-4883-8a5a-b8fe89570d13 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.914606] env[62385]: DEBUG oslo_vmware.api [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 723.914606] env[62385]: value = "task-1205507" [ 723.914606] env[62385]: _type = "Task" [ 723.914606] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.922537] env[62385]: DEBUG oslo_vmware.api [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205507, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.947425] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquiring lock "refresh_cache-f1d33401-35f5-4d79-abb4-26dc6faa784e" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.947771] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquired lock "refresh_cache-f1d33401-35f5-4d79-abb4-26dc6faa784e" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.947843] env[62385]: DEBUG nova.network.neutron [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 724.285754] env[62385]: DEBUG nova.network.neutron [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.350763] env[62385]: DEBUG nova.scheduler.client.report [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 724.428539] env[62385]: DEBUG oslo_vmware.api [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205507, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.509606] env[62385]: DEBUG nova.network.neutron [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 724.729647] env[62385]: DEBUG nova.network.neutron [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Updating instance_info_cache with network_info: [{"id": "65f0aa57-f11b-49fb-8889-1fab7629b3d0", "address": "fa:16:3e:74:d0:58", "network": {"id": "b0a3dea4-1925-4920-8a8e-9dc66733d1da", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1951021425-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ac0c8638f064daba042ded4de8c1cc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee20e439-fed9-490e-97dd-f3c886977ae1", "external-id": "nsx-vlan-transportzone-357", "segmentation_id": 357, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65f0aa57-f1", "ovs_interfaceid": "65f0aa57-f11b-49fb-8889-1fab7629b3d0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.794249] env[62385]: INFO nova.compute.manager [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 866451cc-96c5-433a-a903-9faa1eed538a] Took 1.15 seconds to deallocate network for instance. [ 724.859243] env[62385]: DEBUG oslo_concurrency.lockutils [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.445s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.859775] env[62385]: DEBUG nova.compute.manager [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 724.862428] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 12.859s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.862603] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.862853] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62385) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 724.863122] env[62385]: DEBUG oslo_concurrency.lockutils [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.634s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.864523] env[62385]: INFO nova.compute.claims [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 724.870777] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23bc690a-4bd5-4801-9fe2-dbe425ba3229 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.876655] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3a19bda-bba0-4d4d-917e-5cad69e4b865 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.891160] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7598742-eab1-4ebc-b1fa-15ca1a1016d6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.897424] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf76cc2a-1c81-4a9b-ab29-a46c772beebf {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.932924] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181225MB free_disk=155GB free_vcpus=48 pci_devices=None {{(pid=62385) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 724.933237] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.944978] env[62385]: DEBUG oslo_vmware.api [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205507, 'name': ReconfigVM_Task, 'duration_secs': 0.764652} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.945265] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Reconfigured VM instance instance-0000002c to attach disk [datastore2] 33387505-c576-488b-8c9c-b064fe81a7d7/33387505-c576-488b-8c9c-b064fe81a7d7.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 724.945891] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dea15423-a838-4119-8d4e-34650c87a1ed {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.952921] env[62385]: DEBUG oslo_vmware.api [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 724.952921] env[62385]: value = "task-1205508" [ 724.952921] env[62385]: _type = "Task" [ 724.952921] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.961523] env[62385]: DEBUG oslo_vmware.api [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205508, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.233271] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Releasing lock "refresh_cache-f1d33401-35f5-4d79-abb4-26dc6faa784e" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.233564] env[62385]: DEBUG nova.compute.manager [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Instance network_info: |[{"id": "65f0aa57-f11b-49fb-8889-1fab7629b3d0", "address": "fa:16:3e:74:d0:58", "network": {"id": "b0a3dea4-1925-4920-8a8e-9dc66733d1da", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1951021425-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ac0c8638f064daba042ded4de8c1cc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee20e439-fed9-490e-97dd-f3c886977ae1", "external-id": "nsx-vlan-transportzone-357", "segmentation_id": 357, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65f0aa57-f1", "ovs_interfaceid": "65f0aa57-f11b-49fb-8889-1fab7629b3d0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 725.234023] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:74:d0:58', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ee20e439-fed9-490e-97dd-f3c886977ae1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '65f0aa57-f11b-49fb-8889-1fab7629b3d0', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 725.241709] env[62385]: DEBUG oslo.service.loopingcall [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 725.241917] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 725.242146] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-497005af-79ae-4fad-8ac4-f95f3677b6a9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.261745] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 725.261745] env[62385]: value = "task-1205509" [ 725.261745] env[62385]: _type = "Task" [ 725.261745] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.269145] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205509, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.370151] env[62385]: DEBUG nova.compute.utils [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 725.372324] env[62385]: DEBUG nova.compute.manager [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 725.376676] env[62385]: DEBUG nova.network.neutron [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 725.441562] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0036e774-a1c3-4ab5-928f-e05cc029b599 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Acquiring lock "1ddb4d6d-3872-49db-bb40-e21721241e89" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.441562] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0036e774-a1c3-4ab5-928f-e05cc029b599 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Lock "1ddb4d6d-3872-49db-bb40-e21721241e89" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.441562] env[62385]: DEBUG nova.compute.manager [None req-0036e774-a1c3-4ab5-928f-e05cc029b599 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 725.441562] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9641ab6-0974-461a-b34d-14f2d99cf783 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.449535] env[62385]: DEBUG nova.compute.manager [None req-0036e774-a1c3-4ab5-928f-e05cc029b599 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62385) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 725.450196] env[62385]: DEBUG nova.objects.instance [None req-0036e774-a1c3-4ab5-928f-e05cc029b599 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Lazy-loading 'flavor' on Instance uuid 1ddb4d6d-3872-49db-bb40-e21721241e89 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 725.464492] env[62385]: DEBUG oslo_vmware.api [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205508, 'name': Rename_Task, 'duration_secs': 0.166635} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.464492] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 725.464716] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-71839d56-af22-4612-bea6-d7af99c90a6c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.470966] env[62385]: DEBUG oslo_vmware.api [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 725.470966] env[62385]: value = "task-1205510" [ 725.470966] env[62385]: _type = "Task" [ 725.470966] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.472433] env[62385]: DEBUG nova.policy [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c2ddf01e61874febb9d6fa48751a2b40', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1ff114d055f440e0afc6c87e2ab6b4e9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 725.487469] env[62385]: DEBUG oslo_vmware.api [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205510, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.490672] env[62385]: DEBUG nova.compute.manager [req-1b5fbbe6-f72b-4883-9318-72b8761fb0d5 req-750abe62-bd97-47c1-9b67-0f95d59f9e7f service nova] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Received event network-changed-65f0aa57-f11b-49fb-8889-1fab7629b3d0 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 725.490672] env[62385]: DEBUG nova.compute.manager [req-1b5fbbe6-f72b-4883-9318-72b8761fb0d5 req-750abe62-bd97-47c1-9b67-0f95d59f9e7f service nova] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Refreshing instance network info cache due to event network-changed-65f0aa57-f11b-49fb-8889-1fab7629b3d0. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 725.490672] env[62385]: DEBUG oslo_concurrency.lockutils [req-1b5fbbe6-f72b-4883-9318-72b8761fb0d5 req-750abe62-bd97-47c1-9b67-0f95d59f9e7f service nova] Acquiring lock "refresh_cache-f1d33401-35f5-4d79-abb4-26dc6faa784e" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.490672] env[62385]: DEBUG oslo_concurrency.lockutils [req-1b5fbbe6-f72b-4883-9318-72b8761fb0d5 req-750abe62-bd97-47c1-9b67-0f95d59f9e7f service nova] Acquired lock "refresh_cache-f1d33401-35f5-4d79-abb4-26dc6faa784e" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.490672] env[62385]: DEBUG nova.network.neutron [req-1b5fbbe6-f72b-4883-9318-72b8761fb0d5 req-750abe62-bd97-47c1-9b67-0f95d59f9e7f service nova] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Refreshing network info cache for port 65f0aa57-f11b-49fb-8889-1fab7629b3d0 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 725.772596] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205509, 'name': CreateVM_Task, 'duration_secs': 0.334668} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.772830] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 725.773959] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.773959] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.774181] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 725.774471] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-62ea7aca-cba0-4dd3-a050-d48f9555a91c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.782353] env[62385]: DEBUG oslo_vmware.api [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 725.782353] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]526c8d7a-52f2-acc1-17be-e7331a41a17b" [ 725.782353] env[62385]: _type = "Task" [ 725.782353] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.789118] env[62385]: DEBUG oslo_vmware.api [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]526c8d7a-52f2-acc1-17be-e7331a41a17b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.830130] env[62385]: INFO nova.scheduler.client.report [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Deleted allocations for instance 866451cc-96c5-433a-a903-9faa1eed538a [ 725.882515] env[62385]: DEBUG nova.compute.manager [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 725.895187] env[62385]: DEBUG nova.network.neutron [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Successfully created port: fdae3362-d2d0-4a5e-80ff-c92e442ef9f4 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 725.956741] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0036e774-a1c3-4ab5-928f-e05cc029b599 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 725.958164] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4051a120-39b6-4db8-955b-06c4423c10c6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.966986] env[62385]: DEBUG oslo_vmware.api [None req-0036e774-a1c3-4ab5-928f-e05cc029b599 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Waiting for the task: (returnval){ [ 725.966986] env[62385]: value = "task-1205511" [ 725.966986] env[62385]: _type = "Task" [ 725.966986] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.978042] env[62385]: DEBUG oslo_vmware.api [None req-0036e774-a1c3-4ab5-928f-e05cc029b599 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205511, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.991408] env[62385]: DEBUG oslo_vmware.api [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205510, 'name': PowerOnVM_Task, 'duration_secs': 0.451898} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.991667] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 725.991878] env[62385]: INFO nova.compute.manager [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Took 8.03 seconds to spawn the instance on the hypervisor. [ 725.992096] env[62385]: DEBUG nova.compute.manager [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 725.994833] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44fdcf86-efe7-43f0-a749-d9d6bd56e51d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.293013] env[62385]: DEBUG oslo_vmware.api [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]526c8d7a-52f2-acc1-17be-e7331a41a17b, 'name': SearchDatastore_Task, 'duration_secs': 0.009975} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.293810] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.294100] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 726.294305] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.294451] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.294626] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 726.294890] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-98c23074-552a-4ff9-93a1-9656d6f6602c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.298152] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fd026cf-fe53-48b9-b6b6-08c2ced7cf2e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.305271] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ae0f87c-eaf3-49a4-b8f1-24fa3215d441 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.309862] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 726.310258] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 726.311205] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cebd5fc2-ec3a-4c47-b253-9d1b036eea90 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.342799] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c5796ad3-fe28-40f6-a15e-883540154923 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "866451cc-96c5-433a-a903-9faa1eed538a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 159.908s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.346451] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56fbba44-5fd5-4d21-9f52-de0423f21e55 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.351415] env[62385]: DEBUG oslo_vmware.api [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 726.351415] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]528a885b-76db-1d8c-de4a-b8e218144e0a" [ 726.351415] env[62385]: _type = "Task" [ 726.351415] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.358808] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb14eb4d-147b-4746-833a-f83bc82f8b4c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.365447] env[62385]: DEBUG oslo_vmware.api [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]528a885b-76db-1d8c-de4a-b8e218144e0a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.376598] env[62385]: DEBUG nova.compute.provider_tree [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 726.476189] env[62385]: DEBUG oslo_vmware.api [None req-0036e774-a1c3-4ab5-928f-e05cc029b599 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205511, 'name': PowerOffVM_Task, 'duration_secs': 0.216595} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.476455] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0036e774-a1c3-4ab5-928f-e05cc029b599 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 726.476649] env[62385]: DEBUG nova.compute.manager [None req-0036e774-a1c3-4ab5-928f-e05cc029b599 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 726.478763] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eb2e869-a7c4-4294-809a-912abd85af78 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.493178] env[62385]: DEBUG nova.network.neutron [req-1b5fbbe6-f72b-4883-9318-72b8761fb0d5 req-750abe62-bd97-47c1-9b67-0f95d59f9e7f service nova] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Updated VIF entry in instance network info cache for port 65f0aa57-f11b-49fb-8889-1fab7629b3d0. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 726.493178] env[62385]: DEBUG nova.network.neutron [req-1b5fbbe6-f72b-4883-9318-72b8761fb0d5 req-750abe62-bd97-47c1-9b67-0f95d59f9e7f service nova] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Updating instance_info_cache with network_info: [{"id": "65f0aa57-f11b-49fb-8889-1fab7629b3d0", "address": "fa:16:3e:74:d0:58", "network": {"id": "b0a3dea4-1925-4920-8a8e-9dc66733d1da", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1951021425-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ac0c8638f064daba042ded4de8c1cc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee20e439-fed9-490e-97dd-f3c886977ae1", "external-id": "nsx-vlan-transportzone-357", "segmentation_id": 357, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65f0aa57-f1", "ovs_interfaceid": "65f0aa57-f11b-49fb-8889-1fab7629b3d0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.522488] env[62385]: INFO nova.compute.manager [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Took 30.64 seconds to build instance. [ 726.853040] env[62385]: DEBUG nova.compute.manager [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 726.862964] env[62385]: DEBUG oslo_vmware.api [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]528a885b-76db-1d8c-de4a-b8e218144e0a, 'name': SearchDatastore_Task, 'duration_secs': 0.013274} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.863762] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9969d3f6-7618-484d-a143-c3dae7a36156 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.869246] env[62385]: DEBUG oslo_vmware.api [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 726.869246] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5213f873-19a5-90e5-df76-d1cae8024283" [ 726.869246] env[62385]: _type = "Task" [ 726.869246] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.882825] env[62385]: DEBUG oslo_vmware.api [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5213f873-19a5-90e5-df76-d1cae8024283, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.883642] env[62385]: DEBUG nova.scheduler.client.report [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 726.893889] env[62385]: DEBUG nova.compute.manager [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 726.921732] env[62385]: DEBUG nova.virt.hardware [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 726.921921] env[62385]: DEBUG nova.virt.hardware [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 726.922472] env[62385]: DEBUG nova.virt.hardware [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 726.922748] env[62385]: DEBUG nova.virt.hardware [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 726.922858] env[62385]: DEBUG nova.virt.hardware [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 726.923014] env[62385]: DEBUG nova.virt.hardware [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 726.923645] env[62385]: DEBUG nova.virt.hardware [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 726.923645] env[62385]: DEBUG nova.virt.hardware [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 726.923645] env[62385]: DEBUG nova.virt.hardware [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 726.923874] env[62385]: DEBUG nova.virt.hardware [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 726.923941] env[62385]: DEBUG nova.virt.hardware [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 726.925251] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-650f5555-dd36-48eb-9417-b73f26c1d76b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.936636] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b752db16-c8d4-48a1-bd3c-5a00c2781bfe {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.992282] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0036e774-a1c3-4ab5-928f-e05cc029b599 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Lock "1ddb4d6d-3872-49db-bb40-e21721241e89" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.552s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.997458] env[62385]: DEBUG oslo_concurrency.lockutils [req-1b5fbbe6-f72b-4883-9318-72b8761fb0d5 req-750abe62-bd97-47c1-9b67-0f95d59f9e7f service nova] Releasing lock "refresh_cache-f1d33401-35f5-4d79-abb4-26dc6faa784e" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.997849] env[62385]: DEBUG oslo_concurrency.lockutils [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Acquiring lock "e4089e71-eea1-4e16-b90c-97966333f2bc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.998130] env[62385]: DEBUG oslo_concurrency.lockutils [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Lock "e4089e71-eea1-4e16-b90c-97966333f2bc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 727.024817] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3cc32938-dfb4-4c06-b79e-4f0f5e882849 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "33387505-c576-488b-8c9c-b064fe81a7d7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 152.054s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.380419] env[62385]: DEBUG oslo_vmware.api [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5213f873-19a5-90e5-df76-d1cae8024283, 'name': SearchDatastore_Task, 'duration_secs': 0.012126} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.382435] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.382708] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.383058] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] f1d33401-35f5-4d79-abb4-26dc6faa784e/f1d33401-35f5-4d79-abb4-26dc6faa784e.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 727.386838] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a745cb6d-1d45-454b-b639-a3a7e66635e6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.389037] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Acquiring lock "f2be792c-a723-4a69-b255-bbe7f0e8692d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.389566] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Lock "f2be792c-a723-4a69-b255-bbe7f0e8692d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 727.390151] env[62385]: DEBUG oslo_concurrency.lockutils [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.527s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.390731] env[62385]: DEBUG nova.compute.manager [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 727.393837] env[62385]: DEBUG oslo_concurrency.lockutils [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.131s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 727.394995] env[62385]: INFO nova.compute.claims [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 727.402915] env[62385]: DEBUG oslo_vmware.api [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 727.402915] env[62385]: value = "task-1205512" [ 727.402915] env[62385]: _type = "Task" [ 727.402915] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.411880] env[62385]: DEBUG oslo_vmware.api [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205512, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.529938] env[62385]: DEBUG nova.compute.manager [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 727.901055] env[62385]: DEBUG nova.compute.utils [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 727.906698] env[62385]: DEBUG nova.compute.manager [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 727.909060] env[62385]: DEBUG nova.network.neutron [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 727.925388] env[62385]: DEBUG oslo_vmware.api [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205512, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.988698] env[62385]: DEBUG nova.policy [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8325fcfd0f784303924f518447a76a08', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6bff71f1ab724a50ab687210693ff510', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 728.063711] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.376080] env[62385]: DEBUG nova.network.neutron [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Successfully created port: b40c70f8-444e-4faa-8d57-0455f9d85df1 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 728.409254] env[62385]: DEBUG nova.compute.manager [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 728.427294] env[62385]: DEBUG oslo_vmware.api [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205512, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.452765] env[62385]: DEBUG nova.network.neutron [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Successfully updated port: fdae3362-d2d0-4a5e-80ff-c92e442ef9f4 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 728.577418] env[62385]: DEBUG nova.compute.manager [req-2f8294a7-155d-4703-931f-0c632f2f3728 req-f52b5c08-7b44-4ba8-aad2-fe2b8cea4df1 service nova] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Received event network-vif-plugged-fdae3362-d2d0-4a5e-80ff-c92e442ef9f4 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 728.577679] env[62385]: DEBUG oslo_concurrency.lockutils [req-2f8294a7-155d-4703-931f-0c632f2f3728 req-f52b5c08-7b44-4ba8-aad2-fe2b8cea4df1 service nova] Acquiring lock "f50ddafe-f5a9-422b-b0f4-46c1b111dfbe-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.577964] env[62385]: DEBUG oslo_concurrency.lockutils [req-2f8294a7-155d-4703-931f-0c632f2f3728 req-f52b5c08-7b44-4ba8-aad2-fe2b8cea4df1 service nova] Lock "f50ddafe-f5a9-422b-b0f4-46c1b111dfbe-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.578157] env[62385]: DEBUG oslo_concurrency.lockutils [req-2f8294a7-155d-4703-931f-0c632f2f3728 req-f52b5c08-7b44-4ba8-aad2-fe2b8cea4df1 service nova] Lock "f50ddafe-f5a9-422b-b0f4-46c1b111dfbe-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.578349] env[62385]: DEBUG nova.compute.manager [req-2f8294a7-155d-4703-931f-0c632f2f3728 req-f52b5c08-7b44-4ba8-aad2-fe2b8cea4df1 service nova] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] No waiting events found dispatching network-vif-plugged-fdae3362-d2d0-4a5e-80ff-c92e442ef9f4 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 728.578554] env[62385]: WARNING nova.compute.manager [req-2f8294a7-155d-4703-931f-0c632f2f3728 req-f52b5c08-7b44-4ba8-aad2-fe2b8cea4df1 service nova] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Received unexpected event network-vif-plugged-fdae3362-d2d0-4a5e-80ff-c92e442ef9f4 for instance with vm_state building and task_state spawning. [ 728.664171] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquiring lock "55d62bba-ff30-42bf-b1cb-7567988a9361" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.664171] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "55d62bba-ff30-42bf-b1cb-7567988a9361" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.889643] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2cb5f47-5343-434d-b3b6-d607d789336b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.898697] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f3e369b-e25a-4c3f-abd0-f33fe7302640 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.947034] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0340a52-0b5e-4abf-b14d-6ce96671dc5c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.958044] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32cf173a-e5fa-4077-805d-b798ebe7feaf {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.965656] env[62385]: DEBUG oslo_vmware.api [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205512, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.026012} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.965656] env[62385]: DEBUG oslo_concurrency.lockutils [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Acquiring lock "refresh_cache-f50ddafe-f5a9-422b-b0f4-46c1b111dfbe" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.965656] env[62385]: DEBUG oslo_concurrency.lockutils [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Acquired lock "refresh_cache-f50ddafe-f5a9-422b-b0f4-46c1b111dfbe" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.965656] env[62385]: DEBUG nova.network.neutron [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 728.966240] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] f1d33401-35f5-4d79-abb4-26dc6faa784e/f1d33401-35f5-4d79-abb4-26dc6faa784e.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 728.966240] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 728.966240] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-95b0ee3a-cc1e-4174-a325-8c919dfc930e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.977886] env[62385]: DEBUG nova.compute.provider_tree [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 728.987219] env[62385]: DEBUG oslo_vmware.api [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 728.987219] env[62385]: value = "task-1205513" [ 728.987219] env[62385]: _type = "Task" [ 728.987219] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.998429] env[62385]: DEBUG oslo_vmware.api [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205513, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.453021] env[62385]: DEBUG nova.compute.manager [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 729.482019] env[62385]: DEBUG nova.virt.hardware [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 729.482019] env[62385]: DEBUG nova.virt.hardware [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 729.482019] env[62385]: DEBUG nova.virt.hardware [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 729.482256] env[62385]: DEBUG nova.virt.hardware [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 729.482256] env[62385]: DEBUG nova.virt.hardware [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 729.482256] env[62385]: DEBUG nova.virt.hardware [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 729.482563] env[62385]: DEBUG nova.virt.hardware [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 729.482913] env[62385]: DEBUG nova.virt.hardware [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 729.483193] env[62385]: DEBUG nova.virt.hardware [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 729.483472] env[62385]: DEBUG nova.virt.hardware [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 729.483749] env[62385]: DEBUG nova.virt.hardware [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 729.484664] env[62385]: DEBUG nova.scheduler.client.report [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 729.490829] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24cbc28c-d21d-4e7b-9554-36e9dd4f841b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.505196] env[62385]: DEBUG oslo_vmware.api [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205513, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074817} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.507603] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 729.508865] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e60845a4-c049-4d6a-a7c1-ae027bde43e8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.513148] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccf536d8-7140-4fff-901a-3db041874959 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.521027] env[62385]: DEBUG nova.network.neutron [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 729.554857] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Reconfiguring VM instance instance-0000002d to attach disk [datastore2] f1d33401-35f5-4d79-abb4-26dc6faa784e/f1d33401-35f5-4d79-abb4-26dc6faa784e.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 729.556286] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3de96377-f0cf-40cb-9e7d-473bdb2a8be3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.582028] env[62385]: DEBUG oslo_vmware.api [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 729.582028] env[62385]: value = "task-1205514" [ 729.582028] env[62385]: _type = "Task" [ 729.582028] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.591091] env[62385]: DEBUG oslo_vmware.api [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205514, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.657370] env[62385]: DEBUG nova.objects.instance [None req-936439ba-c825-424e-95e0-fe2f0781199b tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Lazy-loading 'flavor' on Instance uuid 1ddb4d6d-3872-49db-bb40-e21721241e89 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 729.766921] env[62385]: DEBUG nova.network.neutron [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Updating instance_info_cache with network_info: [{"id": "fdae3362-d2d0-4a5e-80ff-c92e442ef9f4", "address": "fa:16:3e:87:99:cf", "network": {"id": "6a072c47-f3cd-4e88-a6f4-bb03318b678f", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1088193534-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ff114d055f440e0afc6c87e2ab6b4e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b5c60ce-845e-4506-bc10-348461fece6d", "external-id": "nsx-vlan-transportzone-831", "segmentation_id": 831, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfdae3362-d2", "ovs_interfaceid": "fdae3362-d2d0-4a5e-80ff-c92e442ef9f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.995571] env[62385]: DEBUG oslo_concurrency.lockutils [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.602s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.996264] env[62385]: DEBUG nova.compute.manager [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 730.000494] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.767s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.005196] env[62385]: INFO nova.compute.claims [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 730.044116] env[62385]: DEBUG nova.compute.manager [req-913f0044-35dc-4558-a019-15a1c2c4658d req-a89a5b72-1cd3-4343-b385-e13da1bf6745 service nova] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Received event network-vif-plugged-b40c70f8-444e-4faa-8d57-0455f9d85df1 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 730.044332] env[62385]: DEBUG oslo_concurrency.lockutils [req-913f0044-35dc-4558-a019-15a1c2c4658d req-a89a5b72-1cd3-4343-b385-e13da1bf6745 service nova] Acquiring lock "8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.046259] env[62385]: DEBUG oslo_concurrency.lockutils [req-913f0044-35dc-4558-a019-15a1c2c4658d req-a89a5b72-1cd3-4343-b385-e13da1bf6745 service nova] Lock "8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.046259] env[62385]: DEBUG oslo_concurrency.lockutils [req-913f0044-35dc-4558-a019-15a1c2c4658d req-a89a5b72-1cd3-4343-b385-e13da1bf6745 service nova] Lock "8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.046259] env[62385]: DEBUG nova.compute.manager [req-913f0044-35dc-4558-a019-15a1c2c4658d req-a89a5b72-1cd3-4343-b385-e13da1bf6745 service nova] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] No waiting events found dispatching network-vif-plugged-b40c70f8-444e-4faa-8d57-0455f9d85df1 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 730.046259] env[62385]: WARNING nova.compute.manager [req-913f0044-35dc-4558-a019-15a1c2c4658d req-a89a5b72-1cd3-4343-b385-e13da1bf6745 service nova] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Received unexpected event network-vif-plugged-b40c70f8-444e-4faa-8d57-0455f9d85df1 for instance with vm_state building and task_state spawning. [ 730.106115] env[62385]: DEBUG oslo_vmware.api [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205514, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.164125] env[62385]: DEBUG oslo_concurrency.lockutils [None req-936439ba-c825-424e-95e0-fe2f0781199b tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Acquiring lock "refresh_cache-1ddb4d6d-3872-49db-bb40-e21721241e89" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.164125] env[62385]: DEBUG oslo_concurrency.lockutils [None req-936439ba-c825-424e-95e0-fe2f0781199b tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Acquired lock "refresh_cache-1ddb4d6d-3872-49db-bb40-e21721241e89" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.164125] env[62385]: DEBUG nova.network.neutron [None req-936439ba-c825-424e-95e0-fe2f0781199b tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 730.164125] env[62385]: DEBUG nova.objects.instance [None req-936439ba-c825-424e-95e0-fe2f0781199b tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Lazy-loading 'info_cache' on Instance uuid 1ddb4d6d-3872-49db-bb40-e21721241e89 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 730.200386] env[62385]: DEBUG nova.network.neutron [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Successfully updated port: b40c70f8-444e-4faa-8d57-0455f9d85df1 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 730.272016] env[62385]: DEBUG oslo_concurrency.lockutils [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Releasing lock "refresh_cache-f50ddafe-f5a9-422b-b0f4-46c1b111dfbe" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.272016] env[62385]: DEBUG nova.compute.manager [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Instance network_info: |[{"id": "fdae3362-d2d0-4a5e-80ff-c92e442ef9f4", "address": "fa:16:3e:87:99:cf", "network": {"id": "6a072c47-f3cd-4e88-a6f4-bb03318b678f", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1088193534-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ff114d055f440e0afc6c87e2ab6b4e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b5c60ce-845e-4506-bc10-348461fece6d", "external-id": "nsx-vlan-transportzone-831", "segmentation_id": 831, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfdae3362-d2", "ovs_interfaceid": "fdae3362-d2d0-4a5e-80ff-c92e442ef9f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 730.272299] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:87:99:cf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4b5c60ce-845e-4506-bc10-348461fece6d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fdae3362-d2d0-4a5e-80ff-c92e442ef9f4', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 730.279629] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Creating folder: Project (1ff114d055f440e0afc6c87e2ab6b4e9). Parent ref: group-v261107. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 730.280705] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ebd48e0a-8e6f-487c-bba6-0f332105e300 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.294131] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Created folder: Project (1ff114d055f440e0afc6c87e2ab6b4e9) in parent group-v261107. [ 730.294131] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Creating folder: Instances. Parent ref: group-v261129. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 730.294131] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a74ad77e-ff68-4815-88bb-b21bae2b9a09 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.306075] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Created folder: Instances in parent group-v261129. [ 730.306075] env[62385]: DEBUG oslo.service.loopingcall [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 730.306075] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 730.306075] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1bd3d3ec-e1d3-4f1e-b302-d9a391a8dd52 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.326244] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 730.326244] env[62385]: value = "task-1205517" [ 730.326244] env[62385]: _type = "Task" [ 730.326244] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.334651] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205517, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.510702] env[62385]: DEBUG nova.compute.utils [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 730.514056] env[62385]: DEBUG nova.compute.manager [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 730.514209] env[62385]: DEBUG nova.network.neutron [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 730.565605] env[62385]: DEBUG nova.policy [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4ac552dc01fc4de2b97e36b41dae725a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '105150c170d74e268e82ab894e9bf0d5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 730.594723] env[62385]: DEBUG oslo_vmware.api [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205514, 'name': ReconfigVM_Task, 'duration_secs': 0.911465} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.594995] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Reconfigured VM instance instance-0000002d to attach disk [datastore2] f1d33401-35f5-4d79-abb4-26dc6faa784e/f1d33401-35f5-4d79-abb4-26dc6faa784e.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 730.595643] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-528cec56-3d8c-4ca3-acaa-6117ef9982e2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.602673] env[62385]: DEBUG oslo_vmware.api [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 730.602673] env[62385]: value = "task-1205518" [ 730.602673] env[62385]: _type = "Task" [ 730.602673] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.607560] env[62385]: DEBUG nova.compute.manager [req-d7199bef-be26-401a-8263-ce5c6377f723 req-55b25190-ca30-4299-8f5b-a074bbd42e59 service nova] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Received event network-changed-fdae3362-d2d0-4a5e-80ff-c92e442ef9f4 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 730.608055] env[62385]: DEBUG nova.compute.manager [req-d7199bef-be26-401a-8263-ce5c6377f723 req-55b25190-ca30-4299-8f5b-a074bbd42e59 service nova] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Refreshing instance network info cache due to event network-changed-fdae3362-d2d0-4a5e-80ff-c92e442ef9f4. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 730.608286] env[62385]: DEBUG oslo_concurrency.lockutils [req-d7199bef-be26-401a-8263-ce5c6377f723 req-55b25190-ca30-4299-8f5b-a074bbd42e59 service nova] Acquiring lock "refresh_cache-f50ddafe-f5a9-422b-b0f4-46c1b111dfbe" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.608443] env[62385]: DEBUG oslo_concurrency.lockutils [req-d7199bef-be26-401a-8263-ce5c6377f723 req-55b25190-ca30-4299-8f5b-a074bbd42e59 service nova] Acquired lock "refresh_cache-f50ddafe-f5a9-422b-b0f4-46c1b111dfbe" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.608614] env[62385]: DEBUG nova.network.neutron [req-d7199bef-be26-401a-8263-ce5c6377f723 req-55b25190-ca30-4299-8f5b-a074bbd42e59 service nova] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Refreshing network info cache for port fdae3362-d2d0-4a5e-80ff-c92e442ef9f4 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 730.615982] env[62385]: DEBUG oslo_vmware.api [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205518, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.667259] env[62385]: DEBUG nova.objects.base [None req-936439ba-c825-424e-95e0-fe2f0781199b tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Object Instance<1ddb4d6d-3872-49db-bb40-e21721241e89> lazy-loaded attributes: flavor,info_cache {{(pid=62385) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 730.703033] env[62385]: DEBUG oslo_concurrency.lockutils [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Acquiring lock "refresh_cache-8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.703213] env[62385]: DEBUG oslo_concurrency.lockutils [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Acquired lock "refresh_cache-8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.703363] env[62385]: DEBUG nova.network.neutron [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 730.838481] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205517, 'name': CreateVM_Task, 'duration_secs': 0.478246} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.838682] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 730.839417] env[62385]: DEBUG oslo_concurrency.lockutils [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.840523] env[62385]: DEBUG oslo_concurrency.lockutils [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.840523] env[62385]: DEBUG oslo_concurrency.lockutils [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 730.840523] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c831d4b-8f93-4bd0-b0ab-dcb31d19e21f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.846939] env[62385]: DEBUG oslo_vmware.api [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Waiting for the task: (returnval){ [ 730.846939] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52d3ff2a-4ad4-6b86-5ea0-6a08814309df" [ 730.846939] env[62385]: _type = "Task" [ 730.846939] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.856114] env[62385]: DEBUG oslo_vmware.api [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52d3ff2a-4ad4-6b86-5ea0-6a08814309df, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.856904] env[62385]: DEBUG nova.network.neutron [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Successfully created port: 35f89d27-3056-4630-b7e1-ee1aa28645b8 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 731.014842] env[62385]: DEBUG nova.compute.manager [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 731.115725] env[62385]: DEBUG oslo_vmware.api [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205518, 'name': Rename_Task, 'duration_secs': 0.18994} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.118172] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 731.118898] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2d81be02-fda4-4c8d-9a4d-13eb4468d308 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.126696] env[62385]: DEBUG oslo_vmware.api [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 731.126696] env[62385]: value = "task-1205519" [ 731.126696] env[62385]: _type = "Task" [ 731.126696] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.139317] env[62385]: DEBUG oslo_vmware.api [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205519, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.284414] env[62385]: DEBUG nova.network.neutron [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 731.340481] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "9b9fd167-d843-4b9d-9e4f-dbde5d3d629e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.340894] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "9b9fd167-d843-4b9d-9e4f-dbde5d3d629e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 731.365081] env[62385]: DEBUG oslo_vmware.api [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52d3ff2a-4ad4-6b86-5ea0-6a08814309df, 'name': SearchDatastore_Task, 'duration_secs': 0.009897} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.365416] env[62385]: DEBUG oslo_concurrency.lockutils [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.365668] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 731.365839] env[62385]: DEBUG oslo_concurrency.lockutils [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 731.366103] env[62385]: DEBUG oslo_concurrency.lockutils [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.366456] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 731.366748] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f0f5e278-f808-4739-8b24-4e8d4af82c4f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.378834] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 731.378834] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 731.379555] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a701fc9-2a29-4712-b3a4-851888955bb4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.389022] env[62385]: DEBUG oslo_vmware.api [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Waiting for the task: (returnval){ [ 731.389022] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521497ba-d5a2-9a7f-5055-53783b1d2021" [ 731.389022] env[62385]: _type = "Task" [ 731.389022] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.398519] env[62385]: DEBUG oslo_vmware.api [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521497ba-d5a2-9a7f-5055-53783b1d2021, 'name': SearchDatastore_Task, 'duration_secs': 0.009633} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.399806] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-21ec4ee1-54d5-4f6f-8c75-8aafb914dd83 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.407423] env[62385]: DEBUG oslo_vmware.api [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Waiting for the task: (returnval){ [ 731.407423] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e29ac3-4027-6c82-1a6d-fbfa402062bd" [ 731.407423] env[62385]: _type = "Task" [ 731.407423] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.416974] env[62385]: DEBUG oslo_vmware.api [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e29ac3-4027-6c82-1a6d-fbfa402062bd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.465465] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5900e17-1914-4bc8-811c-9f346859d6b7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.473742] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-401d81a9-2468-4503-a65e-5b55a705679f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.508505] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef9d1536-abd0-4b21-ac3f-ff46e18cb03a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.517092] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7941150-b6d3-4ca5-afda-38c5a40cd1c1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.537558] env[62385]: DEBUG nova.compute.provider_tree [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 731.638104] env[62385]: DEBUG oslo_vmware.api [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205519, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.795354] env[62385]: DEBUG nova.network.neutron [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Updating instance_info_cache with network_info: [{"id": "b40c70f8-444e-4faa-8d57-0455f9d85df1", "address": "fa:16:3e:54:67:a2", "network": {"id": "f7402a0c-6b85-44f2-800d-4114213ca2c5", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-832196008-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6bff71f1ab724a50ab687210693ff510", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ea0fc1b-0424-46ec-bef5-6b57b7d184d8", "external-id": "nsx-vlan-transportzone-618", "segmentation_id": 618, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb40c70f8-44", "ovs_interfaceid": "b40c70f8-444e-4faa-8d57-0455f9d85df1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.811338] env[62385]: DEBUG nova.network.neutron [req-d7199bef-be26-401a-8263-ce5c6377f723 req-55b25190-ca30-4299-8f5b-a074bbd42e59 service nova] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Updated VIF entry in instance network info cache for port fdae3362-d2d0-4a5e-80ff-c92e442ef9f4. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 731.811709] env[62385]: DEBUG nova.network.neutron [req-d7199bef-be26-401a-8263-ce5c6377f723 req-55b25190-ca30-4299-8f5b-a074bbd42e59 service nova] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Updating instance_info_cache with network_info: [{"id": "fdae3362-d2d0-4a5e-80ff-c92e442ef9f4", "address": "fa:16:3e:87:99:cf", "network": {"id": "6a072c47-f3cd-4e88-a6f4-bb03318b678f", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1088193534-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ff114d055f440e0afc6c87e2ab6b4e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b5c60ce-845e-4506-bc10-348461fece6d", "external-id": "nsx-vlan-transportzone-831", "segmentation_id": 831, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfdae3362-d2", "ovs_interfaceid": "fdae3362-d2d0-4a5e-80ff-c92e442ef9f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.861750] env[62385]: DEBUG nova.network.neutron [None req-936439ba-c825-424e-95e0-fe2f0781199b tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Updating instance_info_cache with network_info: [{"id": "7965d15d-61f5-4fbb-8e38-53e0207a6c58", "address": "fa:16:3e:fb:f0:ae", "network": {"id": "1d0e9098-63d3-4df5-a80a-06f9a5d90db6", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-303269893-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72549cb052314f149111e4e24ffb6656", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "572b7281-aad3-45fa-9cb2-fc1c70569948", "external-id": "nsx-vlan-transportzone-722", "segmentation_id": 722, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7965d15d-61", "ovs_interfaceid": "7965d15d-61f5-4fbb-8e38-53e0207a6c58", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.918209] env[62385]: DEBUG oslo_vmware.api [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e29ac3-4027-6c82-1a6d-fbfa402062bd, 'name': SearchDatastore_Task, 'duration_secs': 0.010974} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.918480] env[62385]: DEBUG oslo_concurrency.lockutils [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.918736] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] f50ddafe-f5a9-422b-b0f4-46c1b111dfbe/f50ddafe-f5a9-422b-b0f4-46c1b111dfbe.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 731.919013] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-38489fa9-94d0-494d-8255-cbe2ccb77769 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.927561] env[62385]: DEBUG oslo_vmware.api [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Waiting for the task: (returnval){ [ 731.927561] env[62385]: value = "task-1205520" [ 731.927561] env[62385]: _type = "Task" [ 731.927561] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.936448] env[62385]: DEBUG oslo_vmware.api [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Task: {'id': task-1205520, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.041120] env[62385]: DEBUG nova.compute.manager [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 732.044143] env[62385]: DEBUG nova.scheduler.client.report [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 732.082765] env[62385]: DEBUG nova.virt.hardware [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 732.082765] env[62385]: DEBUG nova.virt.hardware [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 732.082765] env[62385]: DEBUG nova.virt.hardware [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 732.083690] env[62385]: DEBUG nova.virt.hardware [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 732.083690] env[62385]: DEBUG nova.virt.hardware [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 732.083690] env[62385]: DEBUG nova.virt.hardware [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 732.083948] env[62385]: DEBUG nova.virt.hardware [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 732.084155] env[62385]: DEBUG nova.virt.hardware [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 732.084359] env[62385]: DEBUG nova.virt.hardware [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 732.084547] env[62385]: DEBUG nova.virt.hardware [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 732.084760] env[62385]: DEBUG nova.virt.hardware [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 732.086537] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5e2ab1a-9d1b-49bd-8c25-9680a53ecdb1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.100047] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba1eeac0-da86-40fb-84d9-953572d4c1e2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.143027] env[62385]: DEBUG oslo_vmware.api [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205519, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.153467] env[62385]: DEBUG nova.compute.manager [req-faeb7a43-42e7-4458-acd7-e6340434afb4 req-d8488b1b-813a-4b70-a053-8be1bb8fbb94 service nova] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Received event network-changed-b40c70f8-444e-4faa-8d57-0455f9d85df1 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 732.153671] env[62385]: DEBUG nova.compute.manager [req-faeb7a43-42e7-4458-acd7-e6340434afb4 req-d8488b1b-813a-4b70-a053-8be1bb8fbb94 service nova] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Refreshing instance network info cache due to event network-changed-b40c70f8-444e-4faa-8d57-0455f9d85df1. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 732.153928] env[62385]: DEBUG oslo_concurrency.lockutils [req-faeb7a43-42e7-4458-acd7-e6340434afb4 req-d8488b1b-813a-4b70-a053-8be1bb8fbb94 service nova] Acquiring lock "refresh_cache-8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.300191] env[62385]: DEBUG oslo_concurrency.lockutils [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Releasing lock "refresh_cache-8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.300531] env[62385]: DEBUG nova.compute.manager [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Instance network_info: |[{"id": "b40c70f8-444e-4faa-8d57-0455f9d85df1", "address": "fa:16:3e:54:67:a2", "network": {"id": "f7402a0c-6b85-44f2-800d-4114213ca2c5", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-832196008-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6bff71f1ab724a50ab687210693ff510", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ea0fc1b-0424-46ec-bef5-6b57b7d184d8", "external-id": "nsx-vlan-transportzone-618", "segmentation_id": 618, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb40c70f8-44", "ovs_interfaceid": "b40c70f8-444e-4faa-8d57-0455f9d85df1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 732.300889] env[62385]: DEBUG oslo_concurrency.lockutils [req-faeb7a43-42e7-4458-acd7-e6340434afb4 req-d8488b1b-813a-4b70-a053-8be1bb8fbb94 service nova] Acquired lock "refresh_cache-8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.301073] env[62385]: DEBUG nova.network.neutron [req-faeb7a43-42e7-4458-acd7-e6340434afb4 req-d8488b1b-813a-4b70-a053-8be1bb8fbb94 service nova] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Refreshing network info cache for port b40c70f8-444e-4faa-8d57-0455f9d85df1 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 732.302381] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:54:67:a2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0ea0fc1b-0424-46ec-bef5-6b57b7d184d8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b40c70f8-444e-4faa-8d57-0455f9d85df1', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 732.310614] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Creating folder: Project (6bff71f1ab724a50ab687210693ff510). Parent ref: group-v261107. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 732.311790] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-033a3883-d6a5-42da-9fdd-7d7489d30d0b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.315297] env[62385]: DEBUG oslo_concurrency.lockutils [req-d7199bef-be26-401a-8263-ce5c6377f723 req-55b25190-ca30-4299-8f5b-a074bbd42e59 service nova] Releasing lock "refresh_cache-f50ddafe-f5a9-422b-b0f4-46c1b111dfbe" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.329156] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Created folder: Project (6bff71f1ab724a50ab687210693ff510) in parent group-v261107. [ 732.329879] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Creating folder: Instances. Parent ref: group-v261132. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 732.330626] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7bf77074-9c86-4738-bd45-99f1c56505a5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.344703] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Created folder: Instances in parent group-v261132. [ 732.344703] env[62385]: DEBUG oslo.service.loopingcall [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 732.344900] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 732.344946] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a51ca117-a27f-4faa-b523-55f1b66fb937 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.366794] env[62385]: DEBUG oslo_concurrency.lockutils [None req-936439ba-c825-424e-95e0-fe2f0781199b tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Releasing lock "refresh_cache-1ddb4d6d-3872-49db-bb40-e21721241e89" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.371702] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 732.371702] env[62385]: value = "task-1205523" [ 732.371702] env[62385]: _type = "Task" [ 732.371702] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.381631] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205523, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.436823] env[62385]: DEBUG nova.network.neutron [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Successfully updated port: 35f89d27-3056-4630-b7e1-ee1aa28645b8 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 732.444586] env[62385]: DEBUG oslo_vmware.api [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Task: {'id': task-1205520, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.477384} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.444882] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] f50ddafe-f5a9-422b-b0f4-46c1b111dfbe/f50ddafe-f5a9-422b-b0f4-46c1b111dfbe.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 732.445142] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 732.445434] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-df5c4dc9-0f9f-460c-b996-0248ec4625ff {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.454714] env[62385]: DEBUG oslo_vmware.api [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Waiting for the task: (returnval){ [ 732.454714] env[62385]: value = "task-1205524" [ 732.454714] env[62385]: _type = "Task" [ 732.454714] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.464657] env[62385]: DEBUG oslo_vmware.api [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Task: {'id': task-1205524, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.549899] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.549s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.549899] env[62385]: DEBUG nova.compute.manager [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 732.552445] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.250s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.554032] env[62385]: INFO nova.compute.claims [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 732.641241] env[62385]: DEBUG oslo_vmware.api [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205519, 'name': PowerOnVM_Task, 'duration_secs': 1.285723} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.641574] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 732.641854] env[62385]: INFO nova.compute.manager [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Took 10.18 seconds to spawn the instance on the hypervisor. [ 732.642132] env[62385]: DEBUG nova.compute.manager [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 732.643252] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ecbbe76-4c1a-4546-86c6-387c9a1f16ab {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.870586] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-936439ba-c825-424e-95e0-fe2f0781199b tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 732.870899] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2814b694-c1b8-4294-bd00-82c67c5b6285 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.881858] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205523, 'name': CreateVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.883305] env[62385]: DEBUG oslo_vmware.api [None req-936439ba-c825-424e-95e0-fe2f0781199b tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Waiting for the task: (returnval){ [ 732.883305] env[62385]: value = "task-1205525" [ 732.883305] env[62385]: _type = "Task" [ 732.883305] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.894270] env[62385]: DEBUG oslo_vmware.api [None req-936439ba-c825-424e-95e0-fe2f0781199b tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205525, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.940755] env[62385]: DEBUG oslo_concurrency.lockutils [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "refresh_cache-caca5304-0ad7-427c-812e-de925de63f2c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.940928] env[62385]: DEBUG oslo_concurrency.lockutils [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquired lock "refresh_cache-caca5304-0ad7-427c-812e-de925de63f2c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.941943] env[62385]: DEBUG nova.network.neutron [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 732.968525] env[62385]: DEBUG oslo_vmware.api [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Task: {'id': task-1205524, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065456} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.968814] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 732.969604] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f925ab3-80df-4324-a698-d669e8410e25 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.992666] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Reconfiguring VM instance instance-0000002e to attach disk [datastore2] f50ddafe-f5a9-422b-b0f4-46c1b111dfbe/f50ddafe-f5a9-422b-b0f4-46c1b111dfbe.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 732.993813] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f420ab06-b7f8-4007-9b49-76f6fe2bc02e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.015851] env[62385]: DEBUG oslo_vmware.api [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Waiting for the task: (returnval){ [ 733.015851] env[62385]: value = "task-1205526" [ 733.015851] env[62385]: _type = "Task" [ 733.015851] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.026254] env[62385]: DEBUG oslo_vmware.api [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Task: {'id': task-1205526, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.047188] env[62385]: DEBUG nova.network.neutron [req-faeb7a43-42e7-4458-acd7-e6340434afb4 req-d8488b1b-813a-4b70-a053-8be1bb8fbb94 service nova] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Updated VIF entry in instance network info cache for port b40c70f8-444e-4faa-8d57-0455f9d85df1. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 733.047590] env[62385]: DEBUG nova.network.neutron [req-faeb7a43-42e7-4458-acd7-e6340434afb4 req-d8488b1b-813a-4b70-a053-8be1bb8fbb94 service nova] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Updating instance_info_cache with network_info: [{"id": "b40c70f8-444e-4faa-8d57-0455f9d85df1", "address": "fa:16:3e:54:67:a2", "network": {"id": "f7402a0c-6b85-44f2-800d-4114213ca2c5", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-832196008-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6bff71f1ab724a50ab687210693ff510", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0ea0fc1b-0424-46ec-bef5-6b57b7d184d8", "external-id": "nsx-vlan-transportzone-618", "segmentation_id": 618, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb40c70f8-44", "ovs_interfaceid": "b40c70f8-444e-4faa-8d57-0455f9d85df1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.058059] env[62385]: DEBUG nova.compute.utils [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 733.061815] env[62385]: DEBUG nova.compute.manager [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 733.061987] env[62385]: DEBUG nova.network.neutron [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 733.118761] env[62385]: DEBUG nova.policy [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4a800e72b9e5476fa0e44203f5c7b9ca', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '37c4f2d7d8c24049b1290ec8b0ad9618', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 733.164548] env[62385]: INFO nova.compute.manager [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Took 30.88 seconds to build instance. [ 733.384931] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205523, 'name': CreateVM_Task, 'duration_secs': 0.536601} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.388501] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 733.389264] env[62385]: DEBUG oslo_concurrency.lockutils [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.390224] env[62385]: DEBUG oslo_concurrency.lockutils [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.390342] env[62385]: DEBUG oslo_concurrency.lockutils [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 733.391084] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-64a4e298-72e1-47d1-926d-311b6756e649 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.396910] env[62385]: DEBUG oslo_vmware.api [None req-936439ba-c825-424e-95e0-fe2f0781199b tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205525, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.397710] env[62385]: DEBUG nova.network.neutron [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Successfully created port: 6dd0572e-2124-47c7-b8e5-900aa035f6c4 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 733.403989] env[62385]: DEBUG oslo_vmware.api [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Waiting for the task: (returnval){ [ 733.403989] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c213a1-a4f1-5b2a-0a2c-c609ac9cdf39" [ 733.403989] env[62385]: _type = "Task" [ 733.403989] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.412047] env[62385]: DEBUG oslo_vmware.api [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c213a1-a4f1-5b2a-0a2c-c609ac9cdf39, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.499459] env[62385]: DEBUG nova.network.neutron [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 733.529057] env[62385]: DEBUG oslo_vmware.api [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Task: {'id': task-1205526, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.550618] env[62385]: DEBUG oslo_concurrency.lockutils [req-faeb7a43-42e7-4458-acd7-e6340434afb4 req-d8488b1b-813a-4b70-a053-8be1bb8fbb94 service nova] Releasing lock "refresh_cache-8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.561905] env[62385]: DEBUG nova.compute.manager [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 733.667540] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d043fa4d-d108-4f33-9a01-9f3a4f3f250d tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "f1d33401-35f5-4d79-abb4-26dc6faa784e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 156.947s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.834916] env[62385]: DEBUG nova.network.neutron [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Updating instance_info_cache with network_info: [{"id": "35f89d27-3056-4630-b7e1-ee1aa28645b8", "address": "fa:16:3e:29:b4:dc", "network": {"id": "30c4e78b-6a64-47dc-aaaa-a210160c9f23", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1809587656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "105150c170d74e268e82ab894e9bf0d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10ff2092-e8eb-4768-ad4a-65a80560b447", "external-id": "nsx-vlan-transportzone-845", "segmentation_id": 845, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap35f89d27-30", "ovs_interfaceid": "35f89d27-3056-4630-b7e1-ee1aa28645b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.899653] env[62385]: DEBUG oslo_vmware.api [None req-936439ba-c825-424e-95e0-fe2f0781199b tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205525, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.916538] env[62385]: DEBUG oslo_vmware.api [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c213a1-a4f1-5b2a-0a2c-c609ac9cdf39, 'name': SearchDatastore_Task, 'duration_secs': 0.05331} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.919599] env[62385]: DEBUG oslo_concurrency.lockutils [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.919876] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 733.920174] env[62385]: DEBUG oslo_concurrency.lockutils [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.920323] env[62385]: DEBUG oslo_concurrency.lockutils [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.920500] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 733.921180] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ce74007f-4493-4ef4-9298-d06c1a5f7c25 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.932629] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 733.932820] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 733.936149] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b347b0d4-9151-4f38-9d4f-26cb8fc8a4fb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.942595] env[62385]: DEBUG oslo_vmware.api [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Waiting for the task: (returnval){ [ 733.942595] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52f9041a-c70c-2716-1662-07645fdeed88" [ 733.942595] env[62385]: _type = "Task" [ 733.942595] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.956029] env[62385]: DEBUG oslo_vmware.api [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52f9041a-c70c-2716-1662-07645fdeed88, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.033046] env[62385]: DEBUG oslo_vmware.api [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Task: {'id': task-1205526, 'name': ReconfigVM_Task, 'duration_secs': 0.89712} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.033046] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Reconfigured VM instance instance-0000002e to attach disk [datastore2] f50ddafe-f5a9-422b-b0f4-46c1b111dfbe/f50ddafe-f5a9-422b-b0f4-46c1b111dfbe.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 734.033676] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3d6c5ef0-59bb-4167-8ce4-4d32b12ca86b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.041357] env[62385]: DEBUG oslo_vmware.api [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Waiting for the task: (returnval){ [ 734.041357] env[62385]: value = "task-1205527" [ 734.041357] env[62385]: _type = "Task" [ 734.041357] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.046087] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12aa1a9e-69cf-4be8-b737-23cf16ca14bf {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.056954] env[62385]: DEBUG oslo_vmware.api [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Task: {'id': task-1205527, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.057856] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cc21952-e9b7-440d-a40d-ac0e48294833 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.094978] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ced12c78-9274-4695-bf7d-2dc7499914e4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.104423] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdea64c0-0ab5-4f88-8eac-a5ed871fcad2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.119649] env[62385]: DEBUG nova.compute.provider_tree [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 734.173016] env[62385]: DEBUG nova.compute.manager [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 734.340068] env[62385]: DEBUG oslo_concurrency.lockutils [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Releasing lock "refresh_cache-caca5304-0ad7-427c-812e-de925de63f2c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 734.340439] env[62385]: DEBUG nova.compute.manager [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Instance network_info: |[{"id": "35f89d27-3056-4630-b7e1-ee1aa28645b8", "address": "fa:16:3e:29:b4:dc", "network": {"id": "30c4e78b-6a64-47dc-aaaa-a210160c9f23", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1809587656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "105150c170d74e268e82ab894e9bf0d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10ff2092-e8eb-4768-ad4a-65a80560b447", "external-id": "nsx-vlan-transportzone-845", "segmentation_id": 845, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap35f89d27-30", "ovs_interfaceid": "35f89d27-3056-4630-b7e1-ee1aa28645b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 734.340865] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:29:b4:dc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '10ff2092-e8eb-4768-ad4a-65a80560b447', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '35f89d27-3056-4630-b7e1-ee1aa28645b8', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 734.348525] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Creating folder: Project (105150c170d74e268e82ab894e9bf0d5). Parent ref: group-v261107. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 734.348879] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d1be8b88-1e79-48d7-ae34-d26cc7e69c71 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.361148] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Created folder: Project (105150c170d74e268e82ab894e9bf0d5) in parent group-v261107. [ 734.361407] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Creating folder: Instances. Parent ref: group-v261135. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 734.361722] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-57c7c26e-291f-40b2-bb3a-113f7e49ad5e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.373303] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Created folder: Instances in parent group-v261135. [ 734.373672] env[62385]: DEBUG oslo.service.loopingcall [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 734.373952] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 734.374268] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b613bc1c-ecb1-4e66-978d-c0f0094e274c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.399657] env[62385]: DEBUG oslo_vmware.api [None req-936439ba-c825-424e-95e0-fe2f0781199b tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205525, 'name': PowerOnVM_Task, 'duration_secs': 1.04594} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.403778] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-936439ba-c825-424e-95e0-fe2f0781199b tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 734.403778] env[62385]: DEBUG nova.compute.manager [None req-936439ba-c825-424e-95e0-fe2f0781199b tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 734.403778] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 734.403778] env[62385]: value = "task-1205530" [ 734.403778] env[62385]: _type = "Task" [ 734.403778] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.403778] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4d3ce40-6a8c-42a3-a762-6c4c5376fd94 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.417240] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205530, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.454673] env[62385]: DEBUG oslo_vmware.api [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52f9041a-c70c-2716-1662-07645fdeed88, 'name': SearchDatastore_Task, 'duration_secs': 0.009511} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.455490] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4a0faed6-c028-4431-9d37-8c9a6fd971be {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.461699] env[62385]: DEBUG oslo_vmware.api [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Waiting for the task: (returnval){ [ 734.461699] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52060adb-5cf6-5d03-eb31-55f287c22425" [ 734.461699] env[62385]: _type = "Task" [ 734.461699] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.471455] env[62385]: DEBUG oslo_vmware.api [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52060adb-5cf6-5d03-eb31-55f287c22425, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.531747] env[62385]: DEBUG nova.compute.manager [req-83d349ad-f647-4841-811f-84b408e8b5f7 req-4fb0b3ba-4b10-4958-91f4-de025a43850a service nova] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Received event network-vif-plugged-35f89d27-3056-4630-b7e1-ee1aa28645b8 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 734.531993] env[62385]: DEBUG oslo_concurrency.lockutils [req-83d349ad-f647-4841-811f-84b408e8b5f7 req-4fb0b3ba-4b10-4958-91f4-de025a43850a service nova] Acquiring lock "caca5304-0ad7-427c-812e-de925de63f2c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.532222] env[62385]: DEBUG oslo_concurrency.lockutils [req-83d349ad-f647-4841-811f-84b408e8b5f7 req-4fb0b3ba-4b10-4958-91f4-de025a43850a service nova] Lock "caca5304-0ad7-427c-812e-de925de63f2c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 734.532349] env[62385]: DEBUG oslo_concurrency.lockutils [req-83d349ad-f647-4841-811f-84b408e8b5f7 req-4fb0b3ba-4b10-4958-91f4-de025a43850a service nova] Lock "caca5304-0ad7-427c-812e-de925de63f2c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 734.532865] env[62385]: DEBUG nova.compute.manager [req-83d349ad-f647-4841-811f-84b408e8b5f7 req-4fb0b3ba-4b10-4958-91f4-de025a43850a service nova] [instance: caca5304-0ad7-427c-812e-de925de63f2c] No waiting events found dispatching network-vif-plugged-35f89d27-3056-4630-b7e1-ee1aa28645b8 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 734.532865] env[62385]: WARNING nova.compute.manager [req-83d349ad-f647-4841-811f-84b408e8b5f7 req-4fb0b3ba-4b10-4958-91f4-de025a43850a service nova] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Received unexpected event network-vif-plugged-35f89d27-3056-4630-b7e1-ee1aa28645b8 for instance with vm_state building and task_state spawning. [ 734.533187] env[62385]: DEBUG nova.compute.manager [req-83d349ad-f647-4841-811f-84b408e8b5f7 req-4fb0b3ba-4b10-4958-91f4-de025a43850a service nova] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Received event network-changed-35f89d27-3056-4630-b7e1-ee1aa28645b8 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 734.533357] env[62385]: DEBUG nova.compute.manager [req-83d349ad-f647-4841-811f-84b408e8b5f7 req-4fb0b3ba-4b10-4958-91f4-de025a43850a service nova] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Refreshing instance network info cache due to event network-changed-35f89d27-3056-4630-b7e1-ee1aa28645b8. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 734.533677] env[62385]: DEBUG oslo_concurrency.lockutils [req-83d349ad-f647-4841-811f-84b408e8b5f7 req-4fb0b3ba-4b10-4958-91f4-de025a43850a service nova] Acquiring lock "refresh_cache-caca5304-0ad7-427c-812e-de925de63f2c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.533923] env[62385]: DEBUG oslo_concurrency.lockutils [req-83d349ad-f647-4841-811f-84b408e8b5f7 req-4fb0b3ba-4b10-4958-91f4-de025a43850a service nova] Acquired lock "refresh_cache-caca5304-0ad7-427c-812e-de925de63f2c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.534153] env[62385]: DEBUG nova.network.neutron [req-83d349ad-f647-4841-811f-84b408e8b5f7 req-4fb0b3ba-4b10-4958-91f4-de025a43850a service nova] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Refreshing network info cache for port 35f89d27-3056-4630-b7e1-ee1aa28645b8 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 734.553025] env[62385]: DEBUG oslo_vmware.api [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Task: {'id': task-1205527, 'name': Rename_Task, 'duration_secs': 0.176936} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.553025] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 734.553427] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-07e7e143-f56b-4dad-9fd8-4ae9a73f2ba6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.562234] env[62385]: DEBUG oslo_vmware.api [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Waiting for the task: (returnval){ [ 734.562234] env[62385]: value = "task-1205531" [ 734.562234] env[62385]: _type = "Task" [ 734.562234] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.571619] env[62385]: DEBUG oslo_vmware.api [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Task: {'id': task-1205531, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.599836] env[62385]: DEBUG nova.compute.manager [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 734.622811] env[62385]: DEBUG nova.scheduler.client.report [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 734.631298] env[62385]: DEBUG nova.virt.hardware [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 734.631298] env[62385]: DEBUG nova.virt.hardware [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 734.631437] env[62385]: DEBUG nova.virt.hardware [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 734.631530] env[62385]: DEBUG nova.virt.hardware [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 734.631650] env[62385]: DEBUG nova.virt.hardware [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 734.631801] env[62385]: DEBUG nova.virt.hardware [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 734.632018] env[62385]: DEBUG nova.virt.hardware [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 734.632239] env[62385]: DEBUG nova.virt.hardware [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 734.632345] env[62385]: DEBUG nova.virt.hardware [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 734.632507] env[62385]: DEBUG nova.virt.hardware [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 734.632756] env[62385]: DEBUG nova.virt.hardware [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 734.633720] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3942bfaa-64d8-470a-a2df-49b4b5b021c2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.643411] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efe54c11-89f2-4160-a5d3-81ec0c175b6b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.693362] env[62385]: DEBUG oslo_concurrency.lockutils [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.698541] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquiring lock "679150f5-ad17-45e6-9e3d-9f6e28b1aee3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.698541] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "679150f5-ad17-45e6-9e3d-9f6e28b1aee3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 734.917497] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205530, 'name': CreateVM_Task} progress is 25%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.939281] env[62385]: DEBUG nova.network.neutron [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Successfully updated port: 6dd0572e-2124-47c7-b8e5-900aa035f6c4 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 734.975087] env[62385]: DEBUG oslo_vmware.api [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52060adb-5cf6-5d03-eb31-55f287c22425, 'name': SearchDatastore_Task, 'duration_secs': 0.020521} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.975747] env[62385]: DEBUG oslo_concurrency.lockutils [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 734.976094] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee/8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 734.976409] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8aea930b-b8df-4b4d-8c95-6d8569b0685f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.984809] env[62385]: DEBUG oslo_vmware.api [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Waiting for the task: (returnval){ [ 734.984809] env[62385]: value = "task-1205532" [ 734.984809] env[62385]: _type = "Task" [ 734.984809] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.994638] env[62385]: DEBUG oslo_vmware.api [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Task: {'id': task-1205532, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.073808] env[62385]: DEBUG oslo_vmware.api [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Task: {'id': task-1205531, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.138166] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.585s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 735.138820] env[62385]: DEBUG nova.compute.manager [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 735.143909] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.407s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.145423] env[62385]: INFO nova.compute.claims [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 735.301592] env[62385]: DEBUG nova.network.neutron [req-83d349ad-f647-4841-811f-84b408e8b5f7 req-4fb0b3ba-4b10-4958-91f4-de025a43850a service nova] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Updated VIF entry in instance network info cache for port 35f89d27-3056-4630-b7e1-ee1aa28645b8. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 735.302037] env[62385]: DEBUG nova.network.neutron [req-83d349ad-f647-4841-811f-84b408e8b5f7 req-4fb0b3ba-4b10-4958-91f4-de025a43850a service nova] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Updating instance_info_cache with network_info: [{"id": "35f89d27-3056-4630-b7e1-ee1aa28645b8", "address": "fa:16:3e:29:b4:dc", "network": {"id": "30c4e78b-6a64-47dc-aaaa-a210160c9f23", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1809587656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "105150c170d74e268e82ab894e9bf0d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10ff2092-e8eb-4768-ad4a-65a80560b447", "external-id": "nsx-vlan-transportzone-845", "segmentation_id": 845, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap35f89d27-30", "ovs_interfaceid": "35f89d27-3056-4630-b7e1-ee1aa28645b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.419498] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205530, 'name': CreateVM_Task, 'duration_secs': 0.720056} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.419855] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 735.421014] env[62385]: DEBUG oslo_concurrency.lockutils [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 735.421334] env[62385]: DEBUG oslo_concurrency.lockutils [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.421778] env[62385]: DEBUG oslo_concurrency.lockutils [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 735.422153] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-23628fdc-d4d6-4142-81d8-670a3bb438ed {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.432455] env[62385]: DEBUG oslo_vmware.api [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 735.432455] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]525423d4-d3fc-e1dd-781c-e1f2491a1e21" [ 735.432455] env[62385]: _type = "Task" [ 735.432455] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.445310] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Acquiring lock "refresh_cache-e97ff020-61f3-4947-bb82-5c039ad17747" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 735.445536] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Acquired lock "refresh_cache-e97ff020-61f3-4947-bb82-5c039ad17747" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.445810] env[62385]: DEBUG nova.network.neutron [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 735.447891] env[62385]: DEBUG oslo_vmware.api [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]525423d4-d3fc-e1dd-781c-e1f2491a1e21, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.500593] env[62385]: DEBUG oslo_vmware.api [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Task: {'id': task-1205532, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.573416] env[62385]: DEBUG oslo_vmware.api [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Task: {'id': task-1205531, 'name': PowerOnVM_Task, 'duration_secs': 0.731445} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.573819] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 735.573905] env[62385]: INFO nova.compute.manager [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Took 8.68 seconds to spawn the instance on the hypervisor. [ 735.574105] env[62385]: DEBUG nova.compute.manager [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 735.574911] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-393c5b2d-919b-42d8-951e-993656e19c63 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.652650] env[62385]: DEBUG nova.compute.utils [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 735.657781] env[62385]: DEBUG nova.compute.manager [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 735.657781] env[62385]: DEBUG nova.network.neutron [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 735.699016] env[62385]: DEBUG nova.policy [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4a800e72b9e5476fa0e44203f5c7b9ca', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '37c4f2d7d8c24049b1290ec8b0ad9618', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 735.805249] env[62385]: DEBUG oslo_concurrency.lockutils [req-83d349ad-f647-4841-811f-84b408e8b5f7 req-4fb0b3ba-4b10-4958-91f4-de025a43850a service nova] Releasing lock "refresh_cache-caca5304-0ad7-427c-812e-de925de63f2c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.944016] env[62385]: DEBUG oslo_vmware.api [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]525423d4-d3fc-e1dd-781c-e1f2491a1e21, 'name': SearchDatastore_Task, 'duration_secs': 0.054555} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.944330] env[62385]: DEBUG oslo_concurrency.lockutils [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.944562] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 735.944793] env[62385]: DEBUG oslo_concurrency.lockutils [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 735.944932] env[62385]: DEBUG oslo_concurrency.lockutils [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.945477] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 735.945758] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6ab6aaee-0557-4cf8-990f-61c0375c003a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.958825] env[62385]: DEBUG nova.network.neutron [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Successfully created port: 5a29f755-1606-46ef-9967-30420d891d4f {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 735.961927] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 735.962117] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 735.962840] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b2601b0d-a47c-4b80-96a3-5383910ef3e1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.968507] env[62385]: DEBUG oslo_vmware.api [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 735.968507] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]528ee9a9-3232-46bd-968e-c1f19ab2a475" [ 735.968507] env[62385]: _type = "Task" [ 735.968507] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.979020] env[62385]: DEBUG oslo_vmware.api [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]528ee9a9-3232-46bd-968e-c1f19ab2a475, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.998839] env[62385]: DEBUG oslo_vmware.api [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Task: {'id': task-1205532, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.999951] env[62385]: DEBUG nova.network.neutron [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 736.092846] env[62385]: INFO nova.compute.manager [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Took 29.40 seconds to build instance. [ 736.156066] env[62385]: DEBUG nova.compute.manager [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 736.260874] env[62385]: DEBUG nova.network.neutron [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Updating instance_info_cache with network_info: [{"id": "6dd0572e-2124-47c7-b8e5-900aa035f6c4", "address": "fa:16:3e:d0:62:b9", "network": {"id": "758d5646-566c-4804-bc18-34bd0860fca6", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-677694822-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37c4f2d7d8c24049b1290ec8b0ad9618", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6dd0572e-21", "ovs_interfaceid": "6dd0572e-2124-47c7-b8e5-900aa035f6c4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.484792] env[62385]: DEBUG oslo_vmware.api [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]528ee9a9-3232-46bd-968e-c1f19ab2a475, 'name': SearchDatastore_Task, 'duration_secs': 0.043981} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.484792] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a70824ff-e563-4fe0-8b58-550fd8e6cc32 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.493448] env[62385]: DEBUG oslo_vmware.api [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 736.493448] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52693217-aba5-8514-b743-990352dc6d5a" [ 736.493448] env[62385]: _type = "Task" [ 736.493448] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.502817] env[62385]: DEBUG oslo_vmware.api [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Task: {'id': task-1205532, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.509619] env[62385]: DEBUG oslo_vmware.api [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52693217-aba5-8514-b743-990352dc6d5a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.595066] env[62385]: DEBUG oslo_concurrency.lockutils [None req-73e2cf31-07aa-4188-a1e4-aed441e3531c tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Lock "f50ddafe-f5a9-422b-b0f4-46c1b111dfbe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 157.900s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.638356] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f07f96de-36da-4660-bd75-67eb331db511 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.647672] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-458fcc8f-f0b2-4e46-b112-3e064ba74f07 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.681562] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a04e100-f0ab-4ffb-98d8-70388d45059c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.690144] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3e393bd-1897-4ec6-808c-69bfdb2d5320 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.706666] env[62385]: DEBUG nova.compute.provider_tree [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 736.714958] env[62385]: DEBUG nova.compute.manager [req-49a8c265-268e-408a-afb1-8fd9784a561d req-e54b0707-2367-40c3-976d-2349e1d589f6 service nova] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Received event network-vif-plugged-6dd0572e-2124-47c7-b8e5-900aa035f6c4 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 736.716032] env[62385]: DEBUG oslo_concurrency.lockutils [req-49a8c265-268e-408a-afb1-8fd9784a561d req-e54b0707-2367-40c3-976d-2349e1d589f6 service nova] Acquiring lock "e97ff020-61f3-4947-bb82-5c039ad17747-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.716032] env[62385]: DEBUG oslo_concurrency.lockutils [req-49a8c265-268e-408a-afb1-8fd9784a561d req-e54b0707-2367-40c3-976d-2349e1d589f6 service nova] Lock "e97ff020-61f3-4947-bb82-5c039ad17747-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.716032] env[62385]: DEBUG oslo_concurrency.lockutils [req-49a8c265-268e-408a-afb1-8fd9784a561d req-e54b0707-2367-40c3-976d-2349e1d589f6 service nova] Lock "e97ff020-61f3-4947-bb82-5c039ad17747-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.716032] env[62385]: DEBUG nova.compute.manager [req-49a8c265-268e-408a-afb1-8fd9784a561d req-e54b0707-2367-40c3-976d-2349e1d589f6 service nova] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] No waiting events found dispatching network-vif-plugged-6dd0572e-2124-47c7-b8e5-900aa035f6c4 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 736.716032] env[62385]: WARNING nova.compute.manager [req-49a8c265-268e-408a-afb1-8fd9784a561d req-e54b0707-2367-40c3-976d-2349e1d589f6 service nova] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Received unexpected event network-vif-plugged-6dd0572e-2124-47c7-b8e5-900aa035f6c4 for instance with vm_state building and task_state spawning. [ 736.716431] env[62385]: DEBUG nova.compute.manager [req-49a8c265-268e-408a-afb1-8fd9784a561d req-e54b0707-2367-40c3-976d-2349e1d589f6 service nova] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Received event network-changed-6dd0572e-2124-47c7-b8e5-900aa035f6c4 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 736.716431] env[62385]: DEBUG nova.compute.manager [req-49a8c265-268e-408a-afb1-8fd9784a561d req-e54b0707-2367-40c3-976d-2349e1d589f6 service nova] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Refreshing instance network info cache due to event network-changed-6dd0572e-2124-47c7-b8e5-900aa035f6c4. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 736.716503] env[62385]: DEBUG oslo_concurrency.lockutils [req-49a8c265-268e-408a-afb1-8fd9784a561d req-e54b0707-2367-40c3-976d-2349e1d589f6 service nova] Acquiring lock "refresh_cache-e97ff020-61f3-4947-bb82-5c039ad17747" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.765791] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Releasing lock "refresh_cache-e97ff020-61f3-4947-bb82-5c039ad17747" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 736.766219] env[62385]: DEBUG nova.compute.manager [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Instance network_info: |[{"id": "6dd0572e-2124-47c7-b8e5-900aa035f6c4", "address": "fa:16:3e:d0:62:b9", "network": {"id": "758d5646-566c-4804-bc18-34bd0860fca6", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-677694822-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37c4f2d7d8c24049b1290ec8b0ad9618", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6dd0572e-21", "ovs_interfaceid": "6dd0572e-2124-47c7-b8e5-900aa035f6c4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 736.766582] env[62385]: DEBUG oslo_concurrency.lockutils [req-49a8c265-268e-408a-afb1-8fd9784a561d req-e54b0707-2367-40c3-976d-2349e1d589f6 service nova] Acquired lock "refresh_cache-e97ff020-61f3-4947-bb82-5c039ad17747" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.766776] env[62385]: DEBUG nova.network.neutron [req-49a8c265-268e-408a-afb1-8fd9784a561d req-e54b0707-2367-40c3-976d-2349e1d589f6 service nova] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Refreshing network info cache for port 6dd0572e-2124-47c7-b8e5-900aa035f6c4 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 736.768152] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d0:62:b9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bc1e16db-ad3b-4b7f-ab64-4609c87abac0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6dd0572e-2124-47c7-b8e5-900aa035f6c4', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 736.776800] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Creating folder: Project (37c4f2d7d8c24049b1290ec8b0ad9618). Parent ref: group-v261107. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 736.780447] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-71fb7884-843d-44ae-8874-454c8a5ae298 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.001854] env[62385]: DEBUG oslo_vmware.api [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Task: {'id': task-1205532, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.547124} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.005327] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee/8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 737.005327] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 737.005327] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Created folder: Project (37c4f2d7d8c24049b1290ec8b0ad9618) in parent group-v261107. [ 737.005327] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Creating folder: Instances. Parent ref: group-v261138. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 737.005327] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5535c191-ab9c-4080-a345-802017bfd1f0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.007354] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c6a59655-355d-4870-a826-fcf1c63942bd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.018375] env[62385]: DEBUG oslo_vmware.api [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52693217-aba5-8514-b743-990352dc6d5a, 'name': SearchDatastore_Task, 'duration_secs': 0.050228} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.022253] env[62385]: DEBUG oslo_concurrency.lockutils [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 737.022598] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] caca5304-0ad7-427c-812e-de925de63f2c/caca5304-0ad7-427c-812e-de925de63f2c.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 737.023199] env[62385]: DEBUG oslo_vmware.api [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Waiting for the task: (returnval){ [ 737.023199] env[62385]: value = "task-1205534" [ 737.023199] env[62385]: _type = "Task" [ 737.023199] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.023423] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ee7ef93f-32f8-4caf-a68a-4e1a90ec3fb1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.030562] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Created folder: Instances in parent group-v261138. [ 737.030804] env[62385]: DEBUG oslo.service.loopingcall [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 737.031411] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 737.031957] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-354119e5-4bef-42ff-b615-5bb46d6b8894 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.051823] env[62385]: DEBUG oslo_vmware.api [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Task: {'id': task-1205534, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.052216] env[62385]: DEBUG oslo_vmware.api [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 737.052216] env[62385]: value = "task-1205536" [ 737.052216] env[62385]: _type = "Task" [ 737.052216] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.057917] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 737.057917] env[62385]: value = "task-1205537" [ 737.057917] env[62385]: _type = "Task" [ 737.057917] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.064526] env[62385]: DEBUG oslo_vmware.api [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205536, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.074876] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205537, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.101120] env[62385]: DEBUG nova.compute.manager [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 737.182332] env[62385]: DEBUG nova.compute.manager [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 737.212021] env[62385]: DEBUG nova.scheduler.client.report [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 737.214920] env[62385]: DEBUG nova.virt.hardware [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 737.215280] env[62385]: DEBUG nova.virt.hardware [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 737.215493] env[62385]: DEBUG nova.virt.hardware [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 737.215776] env[62385]: DEBUG nova.virt.hardware [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 737.215977] env[62385]: DEBUG nova.virt.hardware [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 737.216195] env[62385]: DEBUG nova.virt.hardware [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 737.216457] env[62385]: DEBUG nova.virt.hardware [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 737.216667] env[62385]: DEBUG nova.virt.hardware [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 737.216889] env[62385]: DEBUG nova.virt.hardware [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 737.217126] env[62385]: DEBUG nova.virt.hardware [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 737.217383] env[62385]: DEBUG nova.virt.hardware [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 737.218798] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56466419-1922-4671-adf8-037912284ad1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.230300] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bb48cb3-5810-4204-ae6c-b022a55d08a3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.494838] env[62385]: DEBUG nova.network.neutron [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Successfully updated port: 5a29f755-1606-46ef-9967-30420d891d4f {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 737.539719] env[62385]: DEBUG oslo_vmware.api [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Task: {'id': task-1205534, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068895} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.540083] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 737.540985] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fc72f79-79f4-4479-a033-098a62320af5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.566827] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Reconfiguring VM instance instance-0000002f to attach disk [datastore1] 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee/8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 737.572513] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b4614ed6-b350-4b76-89f3-319c0c308a08 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.597349] env[62385]: DEBUG oslo_vmware.api [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205536, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.559153} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.602094] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] caca5304-0ad7-427c-812e-de925de63f2c/caca5304-0ad7-427c-812e-de925de63f2c.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 737.602442] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 737.602727] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205537, 'name': CreateVM_Task, 'duration_secs': 0.377} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.603054] env[62385]: DEBUG oslo_vmware.api [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Waiting for the task: (returnval){ [ 737.603054] env[62385]: value = "task-1205538" [ 737.603054] env[62385]: _type = "Task" [ 737.603054] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.603593] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e838c421-b2d3-40e9-b8e4-c6952e928de4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.605250] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 737.608116] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.608350] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.608748] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 737.618014] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7fdbd713-0e3c-4187-a9a6-17dcbbddd282 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.623598] env[62385]: DEBUG oslo_vmware.api [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Task: {'id': task-1205538, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.623928] env[62385]: DEBUG oslo_vmware.api [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 737.623928] env[62385]: value = "task-1205539" [ 737.623928] env[62385]: _type = "Task" [ 737.623928] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.624929] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Waiting for the task: (returnval){ [ 737.624929] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]523c348b-37c2-2956-74a9-614ef34d2d9d" [ 737.624929] env[62385]: _type = "Task" [ 737.624929] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.633281] env[62385]: DEBUG oslo_concurrency.lockutils [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 737.639253] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]523c348b-37c2-2956-74a9-614ef34d2d9d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.641895] env[62385]: DEBUG oslo_vmware.api [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205539, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.695807] env[62385]: DEBUG nova.network.neutron [req-49a8c265-268e-408a-afb1-8fd9784a561d req-e54b0707-2367-40c3-976d-2349e1d589f6 service nova] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Updated VIF entry in instance network info cache for port 6dd0572e-2124-47c7-b8e5-900aa035f6c4. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 737.696192] env[62385]: DEBUG nova.network.neutron [req-49a8c265-268e-408a-afb1-8fd9784a561d req-e54b0707-2367-40c3-976d-2349e1d589f6 service nova] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Updating instance_info_cache with network_info: [{"id": "6dd0572e-2124-47c7-b8e5-900aa035f6c4", "address": "fa:16:3e:d0:62:b9", "network": {"id": "758d5646-566c-4804-bc18-34bd0860fca6", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-677694822-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37c4f2d7d8c24049b1290ec8b0ad9618", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6dd0572e-21", "ovs_interfaceid": "6dd0572e-2124-47c7-b8e5-900aa035f6c4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.723419] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.579s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.723935] env[62385]: DEBUG nova.compute.manager [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 737.726561] env[62385]: DEBUG oslo_concurrency.lockutils [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.263s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 737.728069] env[62385]: INFO nova.compute.claims [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 737.999068] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Acquiring lock "refresh_cache-749e0976-2e2b-4764-865c-2e630f2edbd1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.999068] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Acquired lock "refresh_cache-749e0976-2e2b-4764-865c-2e630f2edbd1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.999068] env[62385]: DEBUG nova.network.neutron [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 738.116115] env[62385]: DEBUG oslo_vmware.api [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Task: {'id': task-1205538, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.141248] env[62385]: DEBUG oslo_vmware.api [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205539, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063574} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.141832] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]523c348b-37c2-2956-74a9-614ef34d2d9d, 'name': SearchDatastore_Task, 'duration_secs': 0.020317} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.142162] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 738.142447] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.142667] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 738.142893] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.143039] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.143219] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 738.144028] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eb8e151-1647-46af-8371-2c345664ad11 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.146429] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-22d6ce3a-8bb9-4f21-ae1f-d842d6742aa6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.172044] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Reconfiguring VM instance instance-00000030 to attach disk [datastore1] caca5304-0ad7-427c-812e-de925de63f2c/caca5304-0ad7-427c-812e-de925de63f2c.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 738.173384] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b473ef4c-755a-460e-9c5d-1a498b8b162e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.187858] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 738.188078] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 738.188827] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f4f83ab-2658-4e20-aeed-6e227a92a79e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.196045] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Waiting for the task: (returnval){ [ 738.196045] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5216df03-126f-c8cf-8798-fbb2cd36dcda" [ 738.196045] env[62385]: _type = "Task" [ 738.196045] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.200245] env[62385]: DEBUG oslo_concurrency.lockutils [req-49a8c265-268e-408a-afb1-8fd9784a561d req-e54b0707-2367-40c3-976d-2349e1d589f6 service nova] Releasing lock "refresh_cache-e97ff020-61f3-4947-bb82-5c039ad17747" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.200682] env[62385]: DEBUG oslo_vmware.api [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 738.200682] env[62385]: value = "task-1205540" [ 738.200682] env[62385]: _type = "Task" [ 738.200682] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.206877] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5216df03-126f-c8cf-8798-fbb2cd36dcda, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.212798] env[62385]: DEBUG oslo_vmware.api [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205540, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.232426] env[62385]: DEBUG nova.compute.utils [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 738.235717] env[62385]: DEBUG nova.compute.manager [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 738.235875] env[62385]: DEBUG nova.network.neutron [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 738.260879] env[62385]: DEBUG nova.compute.manager [req-f194a19d-d76d-4cce-bcce-3c7d759cfb78 req-cf7844f1-e7b9-4bd3-9ce7-5fd18225afda service nova] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Received event network-changed-fdae3362-d2d0-4a5e-80ff-c92e442ef9f4 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 738.261101] env[62385]: DEBUG nova.compute.manager [req-f194a19d-d76d-4cce-bcce-3c7d759cfb78 req-cf7844f1-e7b9-4bd3-9ce7-5fd18225afda service nova] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Refreshing instance network info cache due to event network-changed-fdae3362-d2d0-4a5e-80ff-c92e442ef9f4. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 738.261320] env[62385]: DEBUG oslo_concurrency.lockutils [req-f194a19d-d76d-4cce-bcce-3c7d759cfb78 req-cf7844f1-e7b9-4bd3-9ce7-5fd18225afda service nova] Acquiring lock "refresh_cache-f50ddafe-f5a9-422b-b0f4-46c1b111dfbe" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.261460] env[62385]: DEBUG oslo_concurrency.lockutils [req-f194a19d-d76d-4cce-bcce-3c7d759cfb78 req-cf7844f1-e7b9-4bd3-9ce7-5fd18225afda service nova] Acquired lock "refresh_cache-f50ddafe-f5a9-422b-b0f4-46c1b111dfbe" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.261615] env[62385]: DEBUG nova.network.neutron [req-f194a19d-d76d-4cce-bcce-3c7d759cfb78 req-cf7844f1-e7b9-4bd3-9ce7-5fd18225afda service nova] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Refreshing network info cache for port fdae3362-d2d0-4a5e-80ff-c92e442ef9f4 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 738.315834] env[62385]: DEBUG nova.policy [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ed4236fea9984babaf0b48de23a3570d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '212a132650584a4d866dc95004faa391', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 738.540533] env[62385]: DEBUG nova.network.neutron [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 738.620839] env[62385]: DEBUG oslo_vmware.api [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Task: {'id': task-1205538, 'name': ReconfigVM_Task, 'duration_secs': 0.897927} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.620839] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Reconfigured VM instance instance-0000002f to attach disk [datastore1] 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee/8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 738.621547] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1423b3de-977a-4a5d-b5cf-13eee75c82c0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.632065] env[62385]: DEBUG oslo_vmware.api [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Waiting for the task: (returnval){ [ 738.632065] env[62385]: value = "task-1205541" [ 738.632065] env[62385]: _type = "Task" [ 738.632065] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.644153] env[62385]: DEBUG oslo_vmware.api [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Task: {'id': task-1205541, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.718618] env[62385]: DEBUG nova.network.neutron [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Updating instance_info_cache with network_info: [{"id": "5a29f755-1606-46ef-9967-30420d891d4f", "address": "fa:16:3e:d8:88:31", "network": {"id": "758d5646-566c-4804-bc18-34bd0860fca6", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-677694822-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37c4f2d7d8c24049b1290ec8b0ad9618", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5a29f755-16", "ovs_interfaceid": "5a29f755-1606-46ef-9967-30420d891d4f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.726794] env[62385]: DEBUG oslo_vmware.api [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205540, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.727087] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5216df03-126f-c8cf-8798-fbb2cd36dcda, 'name': SearchDatastore_Task, 'duration_secs': 0.024566} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.728207] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-95dd7328-3b9e-49b5-92b7-6dd6f8e12dd7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.734739] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Waiting for the task: (returnval){ [ 738.734739] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52dcf00b-b1b5-cbd6-e09d-169e72cc924f" [ 738.734739] env[62385]: _type = "Task" [ 738.734739] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.740163] env[62385]: DEBUG nova.compute.manager [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 738.754905] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52dcf00b-b1b5-cbd6-e09d-169e72cc924f, 'name': SearchDatastore_Task, 'duration_secs': 0.011783} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.755267] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.755679] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] e97ff020-61f3-4947-bb82-5c039ad17747/e97ff020-61f3-4947-bb82-5c039ad17747.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 738.756168] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-449953fc-29e0-4e43-b07e-914bed5ee5e6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.770696] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Waiting for the task: (returnval){ [ 738.770696] env[62385]: value = "task-1205542" [ 738.770696] env[62385]: _type = "Task" [ 738.770696] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.786329] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': task-1205542, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.790403] env[62385]: DEBUG nova.compute.manager [req-eb7e2c3b-ff81-4f5e-b438-a069d9696561 req-94ed2121-dc52-482a-b4d6-e28e7bf7bf7c service nova] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Received event network-vif-plugged-5a29f755-1606-46ef-9967-30420d891d4f {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 738.790628] env[62385]: DEBUG oslo_concurrency.lockutils [req-eb7e2c3b-ff81-4f5e-b438-a069d9696561 req-94ed2121-dc52-482a-b4d6-e28e7bf7bf7c service nova] Acquiring lock "749e0976-2e2b-4764-865c-2e630f2edbd1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.790852] env[62385]: DEBUG oslo_concurrency.lockutils [req-eb7e2c3b-ff81-4f5e-b438-a069d9696561 req-94ed2121-dc52-482a-b4d6-e28e7bf7bf7c service nova] Lock "749e0976-2e2b-4764-865c-2e630f2edbd1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.791152] env[62385]: DEBUG oslo_concurrency.lockutils [req-eb7e2c3b-ff81-4f5e-b438-a069d9696561 req-94ed2121-dc52-482a-b4d6-e28e7bf7bf7c service nova] Lock "749e0976-2e2b-4764-865c-2e630f2edbd1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.791260] env[62385]: DEBUG nova.compute.manager [req-eb7e2c3b-ff81-4f5e-b438-a069d9696561 req-94ed2121-dc52-482a-b4d6-e28e7bf7bf7c service nova] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] No waiting events found dispatching network-vif-plugged-5a29f755-1606-46ef-9967-30420d891d4f {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 738.791419] env[62385]: WARNING nova.compute.manager [req-eb7e2c3b-ff81-4f5e-b438-a069d9696561 req-94ed2121-dc52-482a-b4d6-e28e7bf7bf7c service nova] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Received unexpected event network-vif-plugged-5a29f755-1606-46ef-9967-30420d891d4f for instance with vm_state building and task_state spawning. [ 738.791573] env[62385]: DEBUG nova.compute.manager [req-eb7e2c3b-ff81-4f5e-b438-a069d9696561 req-94ed2121-dc52-482a-b4d6-e28e7bf7bf7c service nova] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Received event network-changed-5a29f755-1606-46ef-9967-30420d891d4f {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 738.791714] env[62385]: DEBUG nova.compute.manager [req-eb7e2c3b-ff81-4f5e-b438-a069d9696561 req-94ed2121-dc52-482a-b4d6-e28e7bf7bf7c service nova] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Refreshing instance network info cache due to event network-changed-5a29f755-1606-46ef-9967-30420d891d4f. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 738.791872] env[62385]: DEBUG oslo_concurrency.lockutils [req-eb7e2c3b-ff81-4f5e-b438-a069d9696561 req-94ed2121-dc52-482a-b4d6-e28e7bf7bf7c service nova] Acquiring lock "refresh_cache-749e0976-2e2b-4764-865c-2e630f2edbd1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.855849] env[62385]: DEBUG nova.network.neutron [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Successfully created port: de1e1c07-3ca9-4e4f-ae07-8e43ed762fc4 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 739.034200] env[62385]: DEBUG nova.network.neutron [req-f194a19d-d76d-4cce-bcce-3c7d759cfb78 req-cf7844f1-e7b9-4bd3-9ce7-5fd18225afda service nova] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Updated VIF entry in instance network info cache for port fdae3362-d2d0-4a5e-80ff-c92e442ef9f4. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 739.034596] env[62385]: DEBUG nova.network.neutron [req-f194a19d-d76d-4cce-bcce-3c7d759cfb78 req-cf7844f1-e7b9-4bd3-9ce7-5fd18225afda service nova] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Updating instance_info_cache with network_info: [{"id": "fdae3362-d2d0-4a5e-80ff-c92e442ef9f4", "address": "fa:16:3e:87:99:cf", "network": {"id": "6a072c47-f3cd-4e88-a6f4-bb03318b678f", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1088193534-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.199", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ff114d055f440e0afc6c87e2ab6b4e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b5c60ce-845e-4506-bc10-348461fece6d", "external-id": "nsx-vlan-transportzone-831", "segmentation_id": 831, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfdae3362-d2", "ovs_interfaceid": "fdae3362-d2d0-4a5e-80ff-c92e442ef9f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.148470] env[62385]: DEBUG oslo_vmware.api [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Task: {'id': task-1205541, 'name': Rename_Task, 'duration_secs': 0.175201} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.148470] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 739.148470] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4603f675-debb-491f-b71e-6141db776d45 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.156926] env[62385]: DEBUG oslo_vmware.api [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Waiting for the task: (returnval){ [ 739.156926] env[62385]: value = "task-1205543" [ 739.156926] env[62385]: _type = "Task" [ 739.156926] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.168697] env[62385]: DEBUG oslo_vmware.api [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Task: {'id': task-1205543, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.199603] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21aca089-124c-4c37-b5d3-d3e8f99fe6f9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.215915] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-165bffe2-fb23-4077-8b1f-58c5027e7749 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.220939] env[62385]: DEBUG oslo_vmware.api [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205540, 'name': ReconfigVM_Task, 'duration_secs': 0.657475} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.220939] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Reconfigured VM instance instance-00000030 to attach disk [datastore1] caca5304-0ad7-427c-812e-de925de63f2c/caca5304-0ad7-427c-812e-de925de63f2c.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 739.222596] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b1e48fa8-8f69-4838-8e59-1150dd6a0be4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.260138] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Releasing lock "refresh_cache-749e0976-2e2b-4764-865c-2e630f2edbd1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.260558] env[62385]: DEBUG nova.compute.manager [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Instance network_info: |[{"id": "5a29f755-1606-46ef-9967-30420d891d4f", "address": "fa:16:3e:d8:88:31", "network": {"id": "758d5646-566c-4804-bc18-34bd0860fca6", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-677694822-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37c4f2d7d8c24049b1290ec8b0ad9618", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5a29f755-16", "ovs_interfaceid": "5a29f755-1606-46ef-9967-30420d891d4f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 739.266035] env[62385]: DEBUG oslo_concurrency.lockutils [req-eb7e2c3b-ff81-4f5e-b438-a069d9696561 req-94ed2121-dc52-482a-b4d6-e28e7bf7bf7c service nova] Acquired lock "refresh_cache-749e0976-2e2b-4764-865c-2e630f2edbd1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.266404] env[62385]: DEBUG nova.network.neutron [req-eb7e2c3b-ff81-4f5e-b438-a069d9696561 req-94ed2121-dc52-482a-b4d6-e28e7bf7bf7c service nova] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Refreshing network info cache for port 5a29f755-1606-46ef-9967-30420d891d4f {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 739.267542] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d8:88:31', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bc1e16db-ad3b-4b7f-ab64-4609c87abac0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5a29f755-1606-46ef-9967-30420d891d4f', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 739.281288] env[62385]: DEBUG oslo.service.loopingcall [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 739.281288] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c92b84f-f876-49c0-b2e2-f48bbaf9821d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.283695] env[62385]: DEBUG oslo_vmware.api [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 739.283695] env[62385]: value = "task-1205544" [ 739.283695] env[62385]: _type = "Task" [ 739.283695] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.287296] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 739.291430] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-268f9d9b-213f-4053-9428-dcc6eae0204d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.320462] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6df27cb-d3c3-43e2-a864-c915e6f615b0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.335849] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': task-1205542, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.337057] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 739.337057] env[62385]: value = "task-1205545" [ 739.337057] env[62385]: _type = "Task" [ 739.337057] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.355074] env[62385]: DEBUG oslo_vmware.api [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205544, 'name': Rename_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.355967] env[62385]: DEBUG nova.compute.provider_tree [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Updating inventory in ProviderTree for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 739.370528] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205545, 'name': CreateVM_Task} progress is 15%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.539846] env[62385]: DEBUG oslo_concurrency.lockutils [req-f194a19d-d76d-4cce-bcce-3c7d759cfb78 req-cf7844f1-e7b9-4bd3-9ce7-5fd18225afda service nova] Releasing lock "refresh_cache-f50ddafe-f5a9-422b-b0f4-46c1b111dfbe" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.561989] env[62385]: DEBUG nova.network.neutron [req-eb7e2c3b-ff81-4f5e-b438-a069d9696561 req-94ed2121-dc52-482a-b4d6-e28e7bf7bf7c service nova] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Updated VIF entry in instance network info cache for port 5a29f755-1606-46ef-9967-30420d891d4f. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 739.562426] env[62385]: DEBUG nova.network.neutron [req-eb7e2c3b-ff81-4f5e-b438-a069d9696561 req-94ed2121-dc52-482a-b4d6-e28e7bf7bf7c service nova] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Updating instance_info_cache with network_info: [{"id": "5a29f755-1606-46ef-9967-30420d891d4f", "address": "fa:16:3e:d8:88:31", "network": {"id": "758d5646-566c-4804-bc18-34bd0860fca6", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-677694822-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37c4f2d7d8c24049b1290ec8b0ad9618", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bc1e16db-ad3b-4b7f-ab64-4609c87abac0", "external-id": "nsx-vlan-transportzone-500", "segmentation_id": 500, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5a29f755-16", "ovs_interfaceid": "5a29f755-1606-46ef-9967-30420d891d4f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.669134] env[62385]: DEBUG oslo_vmware.api [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Task: {'id': task-1205543, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.786365] env[62385]: DEBUG nova.compute.manager [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 739.802094] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': task-1205542, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.725776} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.805472] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] e97ff020-61f3-4947-bb82-5c039ad17747/e97ff020-61f3-4947-bb82-5c039ad17747.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 739.805936] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 739.806088] env[62385]: DEBUG oslo_vmware.api [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205544, 'name': Rename_Task, 'duration_secs': 0.345437} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.808101] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d4f2b130-9556-4881-9078-698aad18fb9f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.809919] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 739.810436] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a23659ea-04d5-417b-8952-114e1df0c940 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.816761] env[62385]: DEBUG nova.virt.hardware [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 739.817057] env[62385]: DEBUG nova.virt.hardware [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 739.817225] env[62385]: DEBUG nova.virt.hardware [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 739.817408] env[62385]: DEBUG nova.virt.hardware [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 739.817553] env[62385]: DEBUG nova.virt.hardware [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 739.817699] env[62385]: DEBUG nova.virt.hardware [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 739.817954] env[62385]: DEBUG nova.virt.hardware [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 739.818167] env[62385]: DEBUG nova.virt.hardware [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 739.818347] env[62385]: DEBUG nova.virt.hardware [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 739.818511] env[62385]: DEBUG nova.virt.hardware [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 739.818687] env[62385]: DEBUG nova.virt.hardware [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 739.819557] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec505583-bf3d-45fd-acb6-5d739c21dfdf {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.824225] env[62385]: DEBUG oslo_vmware.api [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 739.824225] env[62385]: value = "task-1205547" [ 739.824225] env[62385]: _type = "Task" [ 739.824225] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.824495] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Waiting for the task: (returnval){ [ 739.824495] env[62385]: value = "task-1205546" [ 739.824495] env[62385]: _type = "Task" [ 739.824495] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.837088] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c96b964-23c2-4534-99e2-0a023b8e2234 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.847107] env[62385]: DEBUG oslo_vmware.api [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205547, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.847302] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': task-1205546, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.856008] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205545, 'name': CreateVM_Task, 'duration_secs': 0.432358} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.863874] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 739.870060] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.870060] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.870060] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 739.870060] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd3b8a5b-3775-49c5-b665-310bf7456de2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.878380] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Waiting for the task: (returnval){ [ 739.878380] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52809a0b-d390-bde8-091a-a0d874cf40d2" [ 739.878380] env[62385]: _type = "Task" [ 739.878380] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.887075] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52809a0b-d390-bde8-091a-a0d874cf40d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.889308] env[62385]: ERROR nova.scheduler.client.report [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [req-c1526ddf-14f3-4d91-a63d-d705be1b8e53] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-c1526ddf-14f3-4d91-a63d-d705be1b8e53"}]} [ 739.912046] env[62385]: DEBUG nova.scheduler.client.report [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Refreshing inventories for resource provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 739.932582] env[62385]: DEBUG nova.scheduler.client.report [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Updating ProviderTree inventory for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 739.932827] env[62385]: DEBUG nova.compute.provider_tree [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Updating inventory in ProviderTree for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 739.945862] env[62385]: DEBUG nova.scheduler.client.report [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Refreshing aggregate associations for resource provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a, aggregates: None {{(pid=62385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 739.965464] env[62385]: DEBUG nova.scheduler.client.report [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Refreshing trait associations for resource provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=62385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 740.065764] env[62385]: DEBUG oslo_concurrency.lockutils [req-eb7e2c3b-ff81-4f5e-b438-a069d9696561 req-94ed2121-dc52-482a-b4d6-e28e7bf7bf7c service nova] Releasing lock "refresh_cache-749e0976-2e2b-4764-865c-2e630f2edbd1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.168720] env[62385]: DEBUG oslo_vmware.api [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Task: {'id': task-1205543, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.226202] env[62385]: DEBUG oslo_concurrency.lockutils [None req-57e5a242-f0aa-45f3-b3cd-46ad0f3354b0 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Acquiring lock "3e184125-28af-469b-83cc-4ab8859e0e2f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.226440] env[62385]: DEBUG oslo_concurrency.lockutils [None req-57e5a242-f0aa-45f3-b3cd-46ad0f3354b0 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Lock "3e184125-28af-469b-83cc-4ab8859e0e2f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.226640] env[62385]: DEBUG oslo_concurrency.lockutils [None req-57e5a242-f0aa-45f3-b3cd-46ad0f3354b0 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Acquiring lock "3e184125-28af-469b-83cc-4ab8859e0e2f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.226817] env[62385]: DEBUG oslo_concurrency.lockutils [None req-57e5a242-f0aa-45f3-b3cd-46ad0f3354b0 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Lock "3e184125-28af-469b-83cc-4ab8859e0e2f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.226988] env[62385]: DEBUG oslo_concurrency.lockutils [None req-57e5a242-f0aa-45f3-b3cd-46ad0f3354b0 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Lock "3e184125-28af-469b-83cc-4ab8859e0e2f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.229136] env[62385]: INFO nova.compute.manager [None req-57e5a242-f0aa-45f3-b3cd-46ad0f3354b0 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Terminating instance [ 740.230995] env[62385]: DEBUG nova.compute.manager [None req-57e5a242-f0aa-45f3-b3cd-46ad0f3354b0 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 740.231087] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-57e5a242-f0aa-45f3-b3cd-46ad0f3354b0 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 740.231889] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fbc2957-f35b-4103-a55a-a852962fbe21 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.241919] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-57e5a242-f0aa-45f3-b3cd-46ad0f3354b0 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 740.243016] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c0c5af83-9746-4751-ac83-110faaf4c7a2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.249680] env[62385]: DEBUG oslo_vmware.api [None req-57e5a242-f0aa-45f3-b3cd-46ad0f3354b0 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Waiting for the task: (returnval){ [ 740.249680] env[62385]: value = "task-1205548" [ 740.249680] env[62385]: _type = "Task" [ 740.249680] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.261650] env[62385]: DEBUG oslo_vmware.api [None req-57e5a242-f0aa-45f3-b3cd-46ad0f3354b0 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205548, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.341279] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': task-1205546, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.094319} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.344804] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 740.345164] env[62385]: DEBUG oslo_vmware.api [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205547, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.348558] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b606764-2971-4fc7-b8e0-4bfdc708187f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.374353] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Reconfiguring VM instance instance-00000031 to attach disk [datastore1] e97ff020-61f3-4947-bb82-5c039ad17747/e97ff020-61f3-4947-bb82-5c039ad17747.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 740.378714] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-99aea94a-f2f9-4371-8d60-2aa8752b9dee {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.405572] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52809a0b-d390-bde8-091a-a0d874cf40d2, 'name': SearchDatastore_Task, 'duration_secs': 0.014231} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.408605] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.408951] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 740.410030] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.410030] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.410030] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 740.410030] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Waiting for the task: (returnval){ [ 740.410030] env[62385]: value = "task-1205549" [ 740.410030] env[62385]: _type = "Task" [ 740.410030] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.411085] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c98db697-d009-43e1-84d9-bdae14909e3e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.414822] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-56efd9ac-88d1-4cc9-a9f6-f8135e7a519d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.429913] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9600aa9e-ea3d-463f-a148-c5a8f05549ba {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.433479] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': task-1205549, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.434907] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 740.435143] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 740.436328] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5f203b74-fcab-41a0-b23b-7c8adf97abbe {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.467960] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e4c2ae5-425d-4bf8-a81f-e2fc27f49170 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.471530] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Waiting for the task: (returnval){ [ 740.471530] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52802c75-f9e6-bfd9-16c7-c21d2f21949b" [ 740.471530] env[62385]: _type = "Task" [ 740.471530] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.479833] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eea57202-a184-4d19-9c14-70e4767465cf {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.489197] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52802c75-f9e6-bfd9-16c7-c21d2f21949b, 'name': SearchDatastore_Task, 'duration_secs': 0.017245} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.489197] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e7707c03-fc3c-491d-a3aa-ad3cd76cfeb0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.499739] env[62385]: DEBUG nova.compute.provider_tree [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Updating inventory in ProviderTree for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 740.505545] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Waiting for the task: (returnval){ [ 740.505545] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52139c21-907d-6475-e09a-68c9d368c662" [ 740.505545] env[62385]: _type = "Task" [ 740.505545] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.516398] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52139c21-907d-6475-e09a-68c9d368c662, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.640523] env[62385]: DEBUG nova.compute.manager [req-b7b0eca2-278b-4d61-bf7f-81974e0f5fe1 req-a7e8da9f-dae0-44c3-a34b-2186dc14dc88 service nova] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Received event network-vif-plugged-de1e1c07-3ca9-4e4f-ae07-8e43ed762fc4 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 740.640737] env[62385]: DEBUG oslo_concurrency.lockutils [req-b7b0eca2-278b-4d61-bf7f-81974e0f5fe1 req-a7e8da9f-dae0-44c3-a34b-2186dc14dc88 service nova] Acquiring lock "92057af7-28a3-4643-9cda-d3d868d0cfc9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.641056] env[62385]: DEBUG oslo_concurrency.lockutils [req-b7b0eca2-278b-4d61-bf7f-81974e0f5fe1 req-a7e8da9f-dae0-44c3-a34b-2186dc14dc88 service nova] Lock "92057af7-28a3-4643-9cda-d3d868d0cfc9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.641122] env[62385]: DEBUG oslo_concurrency.lockutils [req-b7b0eca2-278b-4d61-bf7f-81974e0f5fe1 req-a7e8da9f-dae0-44c3-a34b-2186dc14dc88 service nova] Lock "92057af7-28a3-4643-9cda-d3d868d0cfc9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.641293] env[62385]: DEBUG nova.compute.manager [req-b7b0eca2-278b-4d61-bf7f-81974e0f5fe1 req-a7e8da9f-dae0-44c3-a34b-2186dc14dc88 service nova] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] No waiting events found dispatching network-vif-plugged-de1e1c07-3ca9-4e4f-ae07-8e43ed762fc4 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 740.641661] env[62385]: WARNING nova.compute.manager [req-b7b0eca2-278b-4d61-bf7f-81974e0f5fe1 req-a7e8da9f-dae0-44c3-a34b-2186dc14dc88 service nova] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Received unexpected event network-vif-plugged-de1e1c07-3ca9-4e4f-ae07-8e43ed762fc4 for instance with vm_state building and task_state spawning. [ 740.668317] env[62385]: DEBUG oslo_vmware.api [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Task: {'id': task-1205543, 'name': PowerOnVM_Task, 'duration_secs': 1.036434} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.668858] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 740.668858] env[62385]: INFO nova.compute.manager [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Took 11.22 seconds to spawn the instance on the hypervisor. [ 740.668960] env[62385]: DEBUG nova.compute.manager [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 740.669773] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d532ace-a2cf-4906-8452-ab75de7be17b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.761428] env[62385]: DEBUG oslo_vmware.api [None req-57e5a242-f0aa-45f3-b3cd-46ad0f3354b0 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205548, 'name': PowerOffVM_Task, 'duration_secs': 0.218217} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.761543] env[62385]: DEBUG nova.network.neutron [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Successfully updated port: de1e1c07-3ca9-4e4f-ae07-8e43ed762fc4 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 740.763184] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-57e5a242-f0aa-45f3-b3cd-46ad0f3354b0 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 740.763184] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-57e5a242-f0aa-45f3-b3cd-46ad0f3354b0 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 740.763416] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-500cc54a-2683-47bf-9a75-cc3a099ae475 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.832545] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-57e5a242-f0aa-45f3-b3cd-46ad0f3354b0 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 740.832813] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-57e5a242-f0aa-45f3-b3cd-46ad0f3354b0 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 740.832967] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-57e5a242-f0aa-45f3-b3cd-46ad0f3354b0 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Deleting the datastore file [datastore1] 3e184125-28af-469b-83cc-4ab8859e0e2f {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 740.833790] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-52212cc3-9031-4a8d-ba8a-9d9c7846b73a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.839113] env[62385]: DEBUG oslo_vmware.api [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205547, 'name': PowerOnVM_Task, 'duration_secs': 0.806959} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.839637] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 740.839841] env[62385]: INFO nova.compute.manager [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Took 8.80 seconds to spawn the instance on the hypervisor. [ 740.840022] env[62385]: DEBUG nova.compute.manager [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 740.840898] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17ad578a-44a4-45cf-b2f7-c134f5f7a5b6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.844362] env[62385]: DEBUG oslo_vmware.api [None req-57e5a242-f0aa-45f3-b3cd-46ad0f3354b0 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Waiting for the task: (returnval){ [ 740.844362] env[62385]: value = "task-1205551" [ 740.844362] env[62385]: _type = "Task" [ 740.844362] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.857618] env[62385]: DEBUG oslo_vmware.api [None req-57e5a242-f0aa-45f3-b3cd-46ad0f3354b0 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205551, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.927356] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': task-1205549, 'name': ReconfigVM_Task, 'duration_secs': 0.347336} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.927609] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Reconfigured VM instance instance-00000031 to attach disk [datastore1] e97ff020-61f3-4947-bb82-5c039ad17747/e97ff020-61f3-4947-bb82-5c039ad17747.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 740.928452] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3b5e24c6-e6e3-4571-aa60-c9ea8816c0e1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.935329] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Waiting for the task: (returnval){ [ 740.935329] env[62385]: value = "task-1205552" [ 740.935329] env[62385]: _type = "Task" [ 740.935329] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.943320] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': task-1205552, 'name': Rename_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.016256] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52139c21-907d-6475-e09a-68c9d368c662, 'name': SearchDatastore_Task, 'duration_secs': 0.027954} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.016898] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 741.016898] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 749e0976-2e2b-4764-865c-2e630f2edbd1/749e0976-2e2b-4764-865c-2e630f2edbd1.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 741.017109] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cd6be0c4-7749-4ac6-a84e-bf2f44c3a0b8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.024912] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Waiting for the task: (returnval){ [ 741.024912] env[62385]: value = "task-1205553" [ 741.024912] env[62385]: _type = "Task" [ 741.024912] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.034101] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': task-1205553, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.045836] env[62385]: DEBUG nova.scheduler.client.report [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Updated inventory for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a with generation 78 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 741.046158] env[62385]: DEBUG nova.compute.provider_tree [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Updating resource provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a generation from 78 to 79 during operation: update_inventory {{(pid=62385) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 741.046349] env[62385]: DEBUG nova.compute.provider_tree [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Updating inventory in ProviderTree for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 741.190206] env[62385]: INFO nova.compute.manager [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Took 24.98 seconds to build instance. [ 741.264106] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Acquiring lock "refresh_cache-92057af7-28a3-4643-9cda-d3d868d0cfc9" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.264274] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Acquired lock "refresh_cache-92057af7-28a3-4643-9cda-d3d868d0cfc9" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.264420] env[62385]: DEBUG nova.network.neutron [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 741.365482] env[62385]: DEBUG oslo_vmware.api [None req-57e5a242-f0aa-45f3-b3cd-46ad0f3354b0 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205551, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.269449} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.368723] env[62385]: INFO nova.compute.manager [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Took 25.12 seconds to build instance. [ 741.370290] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-57e5a242-f0aa-45f3-b3cd-46ad0f3354b0 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 741.370622] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-57e5a242-f0aa-45f3-b3cd-46ad0f3354b0 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 741.370841] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-57e5a242-f0aa-45f3-b3cd-46ad0f3354b0 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 741.371068] env[62385]: INFO nova.compute.manager [None req-57e5a242-f0aa-45f3-b3cd-46ad0f3354b0 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Took 1.14 seconds to destroy the instance on the hypervisor. [ 741.371433] env[62385]: DEBUG oslo.service.loopingcall [None req-57e5a242-f0aa-45f3-b3cd-46ad0f3354b0 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 741.371667] env[62385]: DEBUG nova.compute.manager [-] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 741.371797] env[62385]: DEBUG nova.network.neutron [-] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 741.450531] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': task-1205552, 'name': Rename_Task, 'duration_secs': 0.149913} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.453290] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 741.453984] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c62b56f8-b3f6-419f-a1a9-66feebfbc3a5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.464938] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Waiting for the task: (returnval){ [ 741.464938] env[62385]: value = "task-1205554" [ 741.464938] env[62385]: _type = "Task" [ 741.464938] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.481704] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': task-1205554, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.539453] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': task-1205553, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.552317] env[62385]: DEBUG oslo_concurrency.lockutils [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.825s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.553331] env[62385]: DEBUG nova.compute.manager [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 741.557682] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 16.624s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.692605] env[62385]: DEBUG oslo_concurrency.lockutils [None req-76d7a920-918a-4b1e-8ba8-f83dd51b761e tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Lock "8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 161.596s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.816430] env[62385]: DEBUG nova.network.neutron [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 741.870751] env[62385]: DEBUG oslo_concurrency.lockutils [None req-92d64cac-d242-4ca1-887c-f1194faac44e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "caca5304-0ad7-427c-812e-de925de63f2c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 151.932s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.975853] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': task-1205554, 'name': PowerOnVM_Task} progress is 78%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.036846] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': task-1205553, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.055862] env[62385]: DEBUG nova.network.neutron [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Updating instance_info_cache with network_info: [{"id": "de1e1c07-3ca9-4e4f-ae07-8e43ed762fc4", "address": "fa:16:3e:33:ae:b3", "network": {"id": "3325f6ee-8d27-49f8-8959-af54a4bb2230", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1230956501-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "212a132650584a4d866dc95004faa391", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde1e1c07-3c", "ovs_interfaceid": "de1e1c07-3ca9-4e4f-ae07-8e43ed762fc4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.073116] env[62385]: DEBUG nova.compute.utils [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 742.078160] env[62385]: DEBUG nova.compute.manager [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 742.078160] env[62385]: DEBUG nova.network.neutron [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 742.195310] env[62385]: DEBUG nova.compute.manager [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 742.235904] env[62385]: DEBUG nova.policy [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5e83312d226945c99d05fcc7f74c0978', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bf8e3c942d3445919cfbe988cca84e90', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 742.252557] env[62385]: DEBUG nova.network.neutron [-] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.377613] env[62385]: DEBUG nova.compute.manager [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 742.431753] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8b788b1c-9691-4fa5-a523-a8b477c9025c tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Acquiring lock "8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.432080] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8b788b1c-9691-4fa5-a523-a8b477c9025c tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Lock "8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.432270] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8b788b1c-9691-4fa5-a523-a8b477c9025c tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Acquiring lock "8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.432829] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8b788b1c-9691-4fa5-a523-a8b477c9025c tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Lock "8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.433020] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8b788b1c-9691-4fa5-a523-a8b477c9025c tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Lock "8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.435100] env[62385]: INFO nova.compute.manager [None req-8b788b1c-9691-4fa5-a523-a8b477c9025c tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Terminating instance [ 742.436843] env[62385]: DEBUG nova.compute.manager [None req-8b788b1c-9691-4fa5-a523-a8b477c9025c tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 742.437130] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8b788b1c-9691-4fa5-a523-a8b477c9025c tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 742.438120] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7688c5b-df12-4bac-893a-d8100966d929 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.446132] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b788b1c-9691-4fa5-a523-a8b477c9025c tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 742.446328] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-be026db6-fe59-4d7d-a37b-5e1b9db226c5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.452805] env[62385]: DEBUG oslo_vmware.api [None req-8b788b1c-9691-4fa5-a523-a8b477c9025c tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Waiting for the task: (returnval){ [ 742.452805] env[62385]: value = "task-1205555" [ 742.452805] env[62385]: _type = "Task" [ 742.452805] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.461147] env[62385]: DEBUG oslo_vmware.api [None req-8b788b1c-9691-4fa5-a523-a8b477c9025c tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Task: {'id': task-1205555, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.474984] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': task-1205554, 'name': PowerOnVM_Task, 'duration_secs': 0.778714} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.476380] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 742.476709] env[62385]: INFO nova.compute.manager [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Took 7.88 seconds to spawn the instance on the hypervisor. [ 742.476792] env[62385]: DEBUG nova.compute.manager [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 742.477526] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-802defc1-56ad-4797-ad15-a1248b719974 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.536831] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': task-1205553, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.559011] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Releasing lock "refresh_cache-92057af7-28a3-4643-9cda-d3d868d0cfc9" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.559366] env[62385]: DEBUG nova.compute.manager [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Instance network_info: |[{"id": "de1e1c07-3ca9-4e4f-ae07-8e43ed762fc4", "address": "fa:16:3e:33:ae:b3", "network": {"id": "3325f6ee-8d27-49f8-8959-af54a4bb2230", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1230956501-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "212a132650584a4d866dc95004faa391", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde1e1c07-3c", "ovs_interfaceid": "de1e1c07-3ca9-4e4f-ae07-8e43ed762fc4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 742.559771] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:33:ae:b3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'db68bd64-5b56-49af-a075-13dcf85cb2e0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'de1e1c07-3ca9-4e4f-ae07-8e43ed762fc4', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 742.567690] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Creating folder: Project (212a132650584a4d866dc95004faa391). Parent ref: group-v261107. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 742.567967] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-70826c9c-09eb-4189-8755-c200d2fbb9bd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.575693] env[62385]: DEBUG nova.compute.manager [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 742.584781] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Created folder: Project (212a132650584a4d866dc95004faa391) in parent group-v261107. [ 742.584781] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Creating folder: Instances. Parent ref: group-v261142. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 742.584906] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-29c5bec4-5c9b-4b7e-a35c-e5e9761c7750 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.596664] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Created folder: Instances in parent group-v261142. [ 742.597024] env[62385]: DEBUG oslo.service.loopingcall [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 742.597341] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 742.597341] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-64141914-c523-4e1a-bd25-6fde76ebc019 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.618956] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 1ddb4d6d-3872-49db-bb40-e21721241e89 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 742.619123] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 742.619245] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 3e184125-28af-469b-83cc-4ab8859e0e2f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 742.619360] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 33387505-c576-488b-8c9c-b064fe81a7d7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 742.619470] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance f1d33401-35f5-4d79-abb4-26dc6faa784e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 742.619789] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance f50ddafe-f5a9-422b-b0f4-46c1b111dfbe actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 742.619789] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 742.619789] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance caca5304-0ad7-427c-812e-de925de63f2c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 742.619999] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance e97ff020-61f3-4947-bb82-5c039ad17747 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 742.619999] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 749e0976-2e2b-4764-865c-2e630f2edbd1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 742.620168] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 92057af7-28a3-4643-9cda-d3d868d0cfc9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 742.620292] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 42ecf594-dff7-4af3-ac56-5bbc45b7a192 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 742.624896] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 742.624896] env[62385]: value = "task-1205558" [ 742.624896] env[62385]: _type = "Task" [ 742.624896] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.635696] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205558, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.715098] env[62385]: DEBUG nova.network.neutron [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Successfully created port: 185ea4de-3b36-467d-a112-f439b722671d {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 742.718342] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.755725] env[62385]: INFO nova.compute.manager [-] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Took 1.38 seconds to deallocate network for instance. [ 742.912323] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.963796] env[62385]: DEBUG oslo_vmware.api [None req-8b788b1c-9691-4fa5-a523-a8b477c9025c tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Task: {'id': task-1205555, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.000022] env[62385]: INFO nova.compute.manager [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Took 24.79 seconds to build instance. [ 743.025113] env[62385]: DEBUG nova.compute.manager [req-b2c191f8-dd9b-49bb-95d7-3219fa5709e0 req-77ed4979-83bb-4333-a8da-433155a537c3 service nova] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Received event network-changed-de1e1c07-3ca9-4e4f-ae07-8e43ed762fc4 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 743.025306] env[62385]: DEBUG nova.compute.manager [req-b2c191f8-dd9b-49bb-95d7-3219fa5709e0 req-77ed4979-83bb-4333-a8da-433155a537c3 service nova] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Refreshing instance network info cache due to event network-changed-de1e1c07-3ca9-4e4f-ae07-8e43ed762fc4. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 743.025520] env[62385]: DEBUG oslo_concurrency.lockutils [req-b2c191f8-dd9b-49bb-95d7-3219fa5709e0 req-77ed4979-83bb-4333-a8da-433155a537c3 service nova] Acquiring lock "refresh_cache-92057af7-28a3-4643-9cda-d3d868d0cfc9" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.025656] env[62385]: DEBUG oslo_concurrency.lockutils [req-b2c191f8-dd9b-49bb-95d7-3219fa5709e0 req-77ed4979-83bb-4333-a8da-433155a537c3 service nova] Acquired lock "refresh_cache-92057af7-28a3-4643-9cda-d3d868d0cfc9" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.025807] env[62385]: DEBUG nova.network.neutron [req-b2c191f8-dd9b-49bb-95d7-3219fa5709e0 req-77ed4979-83bb-4333-a8da-433155a537c3 service nova] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Refreshing network info cache for port de1e1c07-3ca9-4e4f-ae07-8e43ed762fc4 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 743.039439] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': task-1205553, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.61701} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.039714] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 749e0976-2e2b-4764-865c-2e630f2edbd1/749e0976-2e2b-4764-865c-2e630f2edbd1.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 743.039929] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 743.040447] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0ea79b01-51ec-4a3b-bff1-0bea31336f78 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.050915] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Waiting for the task: (returnval){ [ 743.050915] env[62385]: value = "task-1205559" [ 743.050915] env[62385]: _type = "Task" [ 743.050915] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.063097] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': task-1205559, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.124828] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance a5c4afc6-38a6-4815-8ec4-cc01c24489bf has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 743.143974] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205558, 'name': CreateVM_Task} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.144213] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 743.144958] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.145127] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.145659] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 743.146342] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1bccd18e-4d95-45c5-82d9-e6c52d7027ea {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.153806] env[62385]: DEBUG oslo_vmware.api [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for the task: (returnval){ [ 743.153806] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a77535-7de9-c83b-c506-a3fdc03a6840" [ 743.153806] env[62385]: _type = "Task" [ 743.153806] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.165948] env[62385]: DEBUG oslo_vmware.api [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a77535-7de9-c83b-c506-a3fdc03a6840, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.267296] env[62385]: DEBUG oslo_concurrency.lockutils [None req-57e5a242-f0aa-45f3-b3cd-46ad0f3354b0 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 743.403463] env[62385]: INFO nova.compute.manager [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Rebuilding instance [ 743.463405] env[62385]: DEBUG nova.compute.manager [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 743.464391] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-113e6fbe-7e72-4cd1-86c9-b9d09dc634a8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.473239] env[62385]: DEBUG oslo_vmware.api [None req-8b788b1c-9691-4fa5-a523-a8b477c9025c tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Task: {'id': task-1205555, 'name': PowerOffVM_Task, 'duration_secs': 0.561548} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.473890] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b788b1c-9691-4fa5-a523-a8b477c9025c tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 743.474138] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8b788b1c-9691-4fa5-a523-a8b477c9025c tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 743.474557] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7e74d525-0365-4011-8e34-a2fe5779082a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.501729] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Lock "e97ff020-61f3-4947-bb82-5c039ad17747" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 147.379s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.546581] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8b788b1c-9691-4fa5-a523-a8b477c9025c tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 743.546770] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8b788b1c-9691-4fa5-a523-a8b477c9025c tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 743.547056] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b788b1c-9691-4fa5-a523-a8b477c9025c tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Deleting the datastore file [datastore1] 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 743.547375] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-729f7a20-dc67-4a07-9e4f-bf2506253360 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.556637] env[62385]: DEBUG oslo_vmware.api [None req-8b788b1c-9691-4fa5-a523-a8b477c9025c tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Waiting for the task: (returnval){ [ 743.556637] env[62385]: value = "task-1205561" [ 743.556637] env[62385]: _type = "Task" [ 743.556637] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.563517] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': task-1205559, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.114467} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.564181] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 743.567151] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a9389cc-c50c-44e9-aa0e-c7205ee075b8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.577124] env[62385]: DEBUG oslo_vmware.api [None req-8b788b1c-9691-4fa5-a523-a8b477c9025c tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Task: {'id': task-1205561, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.596218] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Reconfiguring VM instance instance-00000032 to attach disk [datastore1] 749e0976-2e2b-4764-865c-2e630f2edbd1/749e0976-2e2b-4764-865c-2e630f2edbd1.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 743.597445] env[62385]: DEBUG nova.compute.manager [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 743.604019] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bd88adf7-dea8-46bc-851b-b5bcea83b2b9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.625340] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Waiting for the task: (returnval){ [ 743.625340] env[62385]: value = "task-1205562" [ 743.625340] env[62385]: _type = "Task" [ 743.625340] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.636793] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance e73ecefc-fed3-4ccd-88d3-399a4b72bb4b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 743.636793] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': task-1205562, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.650501] env[62385]: DEBUG nova.virt.hardware [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 743.651015] env[62385]: DEBUG nova.virt.hardware [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 743.651015] env[62385]: DEBUG nova.virt.hardware [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 743.651165] env[62385]: DEBUG nova.virt.hardware [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 743.654975] env[62385]: DEBUG nova.virt.hardware [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 743.654975] env[62385]: DEBUG nova.virt.hardware [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 743.654975] env[62385]: DEBUG nova.virt.hardware [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 743.654975] env[62385]: DEBUG nova.virt.hardware [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 743.654975] env[62385]: DEBUG nova.virt.hardware [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 743.655317] env[62385]: DEBUG nova.virt.hardware [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 743.655317] env[62385]: DEBUG nova.virt.hardware [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 743.655317] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5756bf5e-4b6c-4255-b7ed-e56addb061c7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.674868] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cd16689-0d2a-499b-b92c-b02dbeb57de1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.680074] env[62385]: DEBUG oslo_vmware.api [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a77535-7de9-c83b-c506-a3fdc03a6840, 'name': SearchDatastore_Task, 'duration_secs': 0.016898} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.680448] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.680708] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 743.680936] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.681135] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.681346] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 743.682055] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9b98db4f-17c4-416f-9297-b0ac1655fdb4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.694356] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 743.694565] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 743.695421] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d20118c-2eff-43c3-9e76-0034d55f5145 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.702418] env[62385]: DEBUG oslo_vmware.api [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for the task: (returnval){ [ 743.702418] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52adaf66-2efc-bed3-7029-82500fb19cfe" [ 743.702418] env[62385]: _type = "Task" [ 743.702418] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.725825] env[62385]: DEBUG oslo_vmware.api [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52adaf66-2efc-bed3-7029-82500fb19cfe, 'name': SearchDatastore_Task, 'duration_secs': 0.019849} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.726653] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ad71f0c-19f3-422e-ad5a-e89f3a7341e8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.734252] env[62385]: DEBUG oslo_vmware.api [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for the task: (returnval){ [ 743.734252] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5255b350-20bc-9666-c423-6e10d9fd2bed" [ 743.734252] env[62385]: _type = "Task" [ 743.734252] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.743555] env[62385]: DEBUG oslo_vmware.api [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5255b350-20bc-9666-c423-6e10d9fd2bed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.951998] env[62385]: DEBUG nova.network.neutron [req-b2c191f8-dd9b-49bb-95d7-3219fa5709e0 req-77ed4979-83bb-4333-a8da-433155a537c3 service nova] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Updated VIF entry in instance network info cache for port de1e1c07-3ca9-4e4f-ae07-8e43ed762fc4. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 743.952394] env[62385]: DEBUG nova.network.neutron [req-b2c191f8-dd9b-49bb-95d7-3219fa5709e0 req-77ed4979-83bb-4333-a8da-433155a537c3 service nova] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Updating instance_info_cache with network_info: [{"id": "de1e1c07-3ca9-4e4f-ae07-8e43ed762fc4", "address": "fa:16:3e:33:ae:b3", "network": {"id": "3325f6ee-8d27-49f8-8959-af54a4bb2230", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1230956501-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "212a132650584a4d866dc95004faa391", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapde1e1c07-3c", "ovs_interfaceid": "de1e1c07-3ca9-4e4f-ae07-8e43ed762fc4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.983438] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 743.983677] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-322329e3-5a01-4b2d-82e4-5dc279fc2292 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.991491] env[62385]: DEBUG oslo_vmware.api [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 743.991491] env[62385]: value = "task-1205563" [ 743.991491] env[62385]: _type = "Task" [ 743.991491] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.001906] env[62385]: DEBUG oslo_vmware.api [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205563, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.004403] env[62385]: DEBUG nova.compute.manager [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 744.069041] env[62385]: DEBUG oslo_vmware.api [None req-8b788b1c-9691-4fa5-a523-a8b477c9025c tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Task: {'id': task-1205561, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.259601} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.069358] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b788b1c-9691-4fa5-a523-a8b477c9025c tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 744.069577] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8b788b1c-9691-4fa5-a523-a8b477c9025c tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 744.069788] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8b788b1c-9691-4fa5-a523-a8b477c9025c tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 744.070041] env[62385]: INFO nova.compute.manager [None req-8b788b1c-9691-4fa5-a523-a8b477c9025c tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Took 1.63 seconds to destroy the instance on the hypervisor. [ 744.070355] env[62385]: DEBUG oslo.service.loopingcall [None req-8b788b1c-9691-4fa5-a523-a8b477c9025c tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 744.070627] env[62385]: DEBUG nova.compute.manager [-] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 744.070668] env[62385]: DEBUG nova.network.neutron [-] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 744.136761] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': task-1205562, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.139845] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 65bb1d14-84b2-4e75-acdc-dc674a035101 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 744.245356] env[62385]: DEBUG oslo_vmware.api [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5255b350-20bc-9666-c423-6e10d9fd2bed, 'name': SearchDatastore_Task, 'duration_secs': 0.011342} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.245570] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.245820] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 92057af7-28a3-4643-9cda-d3d868d0cfc9/92057af7-28a3-4643-9cda-d3d868d0cfc9.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 744.246084] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ca3aac0f-f11c-4b73-8273-7dbf963ece58 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.254314] env[62385]: DEBUG oslo_vmware.api [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for the task: (returnval){ [ 744.254314] env[62385]: value = "task-1205564" [ 744.254314] env[62385]: _type = "Task" [ 744.254314] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.263172] env[62385]: DEBUG oslo_vmware.api [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205564, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.406352] env[62385]: DEBUG nova.compute.manager [req-4b30f4a0-d6d9-4abf-8684-2b48d9549d53 req-21877056-909d-4ed3-bb13-aaa2a7f2f12d service nova] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Received event network-vif-deleted-b40c70f8-444e-4faa-8d57-0455f9d85df1 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 744.406407] env[62385]: INFO nova.compute.manager [req-4b30f4a0-d6d9-4abf-8684-2b48d9549d53 req-21877056-909d-4ed3-bb13-aaa2a7f2f12d service nova] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Neutron deleted interface b40c70f8-444e-4faa-8d57-0455f9d85df1; detaching it from the instance and deleting it from the info cache [ 744.406578] env[62385]: DEBUG nova.network.neutron [req-4b30f4a0-d6d9-4abf-8684-2b48d9549d53 req-21877056-909d-4ed3-bb13-aaa2a7f2f12d service nova] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.455314] env[62385]: DEBUG oslo_concurrency.lockutils [req-b2c191f8-dd9b-49bb-95d7-3219fa5709e0 req-77ed4979-83bb-4333-a8da-433155a537c3 service nova] Releasing lock "refresh_cache-92057af7-28a3-4643-9cda-d3d868d0cfc9" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.455592] env[62385]: DEBUG nova.compute.manager [req-b2c191f8-dd9b-49bb-95d7-3219fa5709e0 req-77ed4979-83bb-4333-a8da-433155a537c3 service nova] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Received event network-vif-deleted-60df554a-a3e2-41eb-b893-16459c19bc7c {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 744.505842] env[62385]: DEBUG oslo_vmware.api [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205563, 'name': PowerOffVM_Task, 'duration_secs': 0.200947} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.505842] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 744.505842] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 744.507281] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f89e670d-48aa-4d71-b739-e6ac2795cc22 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.521807] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 744.521807] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-20c5a1fe-f4a8-4122-9119-251726f2d0e4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.534729] env[62385]: DEBUG oslo_concurrency.lockutils [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.600693] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 744.600985] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 744.601242] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Deleting the datastore file [datastore1] caca5304-0ad7-427c-812e-de925de63f2c {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 744.601560] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-71e38100-6a0c-47db-a576-98215e47bbf9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.610556] env[62385]: DEBUG oslo_vmware.api [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 744.610556] env[62385]: value = "task-1205566" [ 744.610556] env[62385]: _type = "Task" [ 744.610556] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.621848] env[62385]: DEBUG oslo_vmware.api [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205566, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.636980] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': task-1205562, 'name': ReconfigVM_Task, 'duration_secs': 0.562605} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.637862] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Reconfigured VM instance instance-00000032 to attach disk [datastore1] 749e0976-2e2b-4764-865c-2e630f2edbd1/749e0976-2e2b-4764-865c-2e630f2edbd1.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 744.638585] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-231221b3-09d0-4714-ba29-be7969280521 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.643092] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance f8a98710-9d02-4aa3-b6a5-34b7060ad62d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 744.650387] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Waiting for the task: (returnval){ [ 744.650387] env[62385]: value = "task-1205567" [ 744.650387] env[62385]: _type = "Task" [ 744.650387] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.663184] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': task-1205567, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.767764] env[62385]: DEBUG oslo_vmware.api [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205564, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.498377} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.767764] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 92057af7-28a3-4643-9cda-d3d868d0cfc9/92057af7-28a3-4643-9cda-d3d868d0cfc9.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 744.767764] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 744.767764] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-adda7e45-1c67-42da-9604-b45c8bb77575 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.775027] env[62385]: DEBUG oslo_vmware.api [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for the task: (returnval){ [ 744.775027] env[62385]: value = "task-1205568" [ 744.775027] env[62385]: _type = "Task" [ 744.775027] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.787667] env[62385]: DEBUG oslo_vmware.api [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205568, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.808485] env[62385]: DEBUG nova.network.neutron [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Successfully updated port: 185ea4de-3b36-467d-a112-f439b722671d {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 744.842739] env[62385]: DEBUG nova.network.neutron [-] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.909433] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-32777afe-89f4-413f-b064-3495ea618563 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.919933] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7ef1ae5-05a9-4e99-a73b-c0db9e1c3de6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.953058] env[62385]: DEBUG nova.compute.manager [req-4b30f4a0-d6d9-4abf-8684-2b48d9549d53 req-21877056-909d-4ed3-bb13-aaa2a7f2f12d service nova] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Detach interface failed, port_id=b40c70f8-444e-4faa-8d57-0455f9d85df1, reason: Instance 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 745.088429] env[62385]: DEBUG nova.compute.manager [req-e5513ad5-7a17-4990-b7d0-adebb47cf2f8 req-d3a867a0-5379-493d-9def-35fbbe3e9566 service nova] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Received event network-vif-plugged-185ea4de-3b36-467d-a112-f439b722671d {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 745.088641] env[62385]: DEBUG oslo_concurrency.lockutils [req-e5513ad5-7a17-4990-b7d0-adebb47cf2f8 req-d3a867a0-5379-493d-9def-35fbbe3e9566 service nova] Acquiring lock "42ecf594-dff7-4af3-ac56-5bbc45b7a192-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 745.088868] env[62385]: DEBUG oslo_concurrency.lockutils [req-e5513ad5-7a17-4990-b7d0-adebb47cf2f8 req-d3a867a0-5379-493d-9def-35fbbe3e9566 service nova] Lock "42ecf594-dff7-4af3-ac56-5bbc45b7a192-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 745.089117] env[62385]: DEBUG oslo_concurrency.lockutils [req-e5513ad5-7a17-4990-b7d0-adebb47cf2f8 req-d3a867a0-5379-493d-9def-35fbbe3e9566 service nova] Lock "42ecf594-dff7-4af3-ac56-5bbc45b7a192-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.089201] env[62385]: DEBUG nova.compute.manager [req-e5513ad5-7a17-4990-b7d0-adebb47cf2f8 req-d3a867a0-5379-493d-9def-35fbbe3e9566 service nova] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] No waiting events found dispatching network-vif-plugged-185ea4de-3b36-467d-a112-f439b722671d {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 745.089382] env[62385]: WARNING nova.compute.manager [req-e5513ad5-7a17-4990-b7d0-adebb47cf2f8 req-d3a867a0-5379-493d-9def-35fbbe3e9566 service nova] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Received unexpected event network-vif-plugged-185ea4de-3b36-467d-a112-f439b722671d for instance with vm_state building and task_state spawning. [ 745.089526] env[62385]: DEBUG nova.compute.manager [req-e5513ad5-7a17-4990-b7d0-adebb47cf2f8 req-d3a867a0-5379-493d-9def-35fbbe3e9566 service nova] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Received event network-changed-185ea4de-3b36-467d-a112-f439b722671d {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 745.089641] env[62385]: DEBUG nova.compute.manager [req-e5513ad5-7a17-4990-b7d0-adebb47cf2f8 req-d3a867a0-5379-493d-9def-35fbbe3e9566 service nova] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Refreshing instance network info cache due to event network-changed-185ea4de-3b36-467d-a112-f439b722671d. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 745.089816] env[62385]: DEBUG oslo_concurrency.lockutils [req-e5513ad5-7a17-4990-b7d0-adebb47cf2f8 req-d3a867a0-5379-493d-9def-35fbbe3e9566 service nova] Acquiring lock "refresh_cache-42ecf594-dff7-4af3-ac56-5bbc45b7a192" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.089947] env[62385]: DEBUG oslo_concurrency.lockutils [req-e5513ad5-7a17-4990-b7d0-adebb47cf2f8 req-d3a867a0-5379-493d-9def-35fbbe3e9566 service nova] Acquired lock "refresh_cache-42ecf594-dff7-4af3-ac56-5bbc45b7a192" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.090139] env[62385]: DEBUG nova.network.neutron [req-e5513ad5-7a17-4990-b7d0-adebb47cf2f8 req-d3a867a0-5379-493d-9def-35fbbe3e9566 service nova] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Refreshing network info cache for port 185ea4de-3b36-467d-a112-f439b722671d {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 745.120413] env[62385]: DEBUG oslo_vmware.api [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205566, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.226258} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.120675] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 745.120832] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 745.121012] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 745.146949] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance f77dc5ab-7e63-4a1b-9c87-e474a7e28c93 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 745.160611] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': task-1205567, 'name': Rename_Task, 'duration_secs': 0.189657} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.161107] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 745.161234] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5117d37c-a223-4364-ab21-1e23d6118c46 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.168858] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Waiting for the task: (returnval){ [ 745.168858] env[62385]: value = "task-1205569" [ 745.168858] env[62385]: _type = "Task" [ 745.168858] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.177913] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': task-1205569, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.283043] env[62385]: DEBUG oslo_vmware.api [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205568, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078492} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.283334] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 745.284119] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22e625ad-55ea-415f-b33d-3bf00e36f00f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.306858] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Reconfiguring VM instance instance-00000033 to attach disk [datastore2] 92057af7-28a3-4643-9cda-d3d868d0cfc9/92057af7-28a3-4643-9cda-d3d868d0cfc9.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 745.307146] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5ce083bb-78bf-4249-a4c5-36fb3a30eef6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.321468] env[62385]: DEBUG oslo_concurrency.lockutils [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "refresh_cache-42ecf594-dff7-4af3-ac56-5bbc45b7a192" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.327607] env[62385]: DEBUG oslo_vmware.api [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for the task: (returnval){ [ 745.327607] env[62385]: value = "task-1205570" [ 745.327607] env[62385]: _type = "Task" [ 745.327607] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.335624] env[62385]: DEBUG oslo_vmware.api [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205570, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.345295] env[62385]: INFO nova.compute.manager [-] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Took 1.27 seconds to deallocate network for instance. [ 745.622252] env[62385]: DEBUG nova.network.neutron [req-e5513ad5-7a17-4990-b7d0-adebb47cf2f8 req-d3a867a0-5379-493d-9def-35fbbe3e9566 service nova] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 745.650576] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 745.682780] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': task-1205569, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.744350] env[62385]: DEBUG nova.network.neutron [req-e5513ad5-7a17-4990-b7d0-adebb47cf2f8 req-d3a867a0-5379-493d-9def-35fbbe3e9566 service nova] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.838066] env[62385]: DEBUG oslo_vmware.api [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205570, 'name': ReconfigVM_Task, 'duration_secs': 0.447234} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.838366] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Reconfigured VM instance instance-00000033 to attach disk [datastore2] 92057af7-28a3-4643-9cda-d3d868d0cfc9/92057af7-28a3-4643-9cda-d3d868d0cfc9.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 745.838989] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-546d53a1-9581-4b10-a13f-eff668a98199 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.845574] env[62385]: DEBUG oslo_vmware.api [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for the task: (returnval){ [ 745.845574] env[62385]: value = "task-1205571" [ 745.845574] env[62385]: _type = "Task" [ 745.845574] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.851556] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8b788b1c-9691-4fa5-a523-a8b477c9025c tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 745.854562] env[62385]: DEBUG oslo_vmware.api [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205571, 'name': Rename_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.154095] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 0d823fcc-cbba-41f6-a47c-2f57d3e79948 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 746.157777] env[62385]: DEBUG nova.virt.hardware [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 746.158329] env[62385]: DEBUG nova.virt.hardware [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 746.158606] env[62385]: DEBUG nova.virt.hardware [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 746.158996] env[62385]: DEBUG nova.virt.hardware [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 746.159268] env[62385]: DEBUG nova.virt.hardware [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 746.159472] env[62385]: DEBUG nova.virt.hardware [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 746.159728] env[62385]: DEBUG nova.virt.hardware [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 746.159926] env[62385]: DEBUG nova.virt.hardware [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 746.160209] env[62385]: DEBUG nova.virt.hardware [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 746.161031] env[62385]: DEBUG nova.virt.hardware [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 746.161031] env[62385]: DEBUG nova.virt.hardware [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 746.161568] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a31f4101-d243-4d69-83dd-36ca82e4039f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.174569] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d04d86e-2c31-4bcc-9f39-8a83896586c4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.192412] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': task-1205569, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.193251] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:29:b4:dc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '10ff2092-e8eb-4768-ad4a-65a80560b447', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '35f89d27-3056-4630-b7e1-ee1aa28645b8', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 746.201829] env[62385]: DEBUG oslo.service.loopingcall [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 746.202108] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 746.202333] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-18faab8b-f758-4a11-be7e-ee8406f9f377 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.224176] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 746.224176] env[62385]: value = "task-1205572" [ 746.224176] env[62385]: _type = "Task" [ 746.224176] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.232954] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205572, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.246618] env[62385]: DEBUG oslo_concurrency.lockutils [req-e5513ad5-7a17-4990-b7d0-adebb47cf2f8 req-d3a867a0-5379-493d-9def-35fbbe3e9566 service nova] Releasing lock "refresh_cache-42ecf594-dff7-4af3-ac56-5bbc45b7a192" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.247011] env[62385]: DEBUG oslo_concurrency.lockutils [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquired lock "refresh_cache-42ecf594-dff7-4af3-ac56-5bbc45b7a192" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.247178] env[62385]: DEBUG nova.network.neutron [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 746.356105] env[62385]: DEBUG oslo_vmware.api [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205571, 'name': Rename_Task, 'duration_secs': 0.171378} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.356407] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 746.359254] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-221c35cd-66b4-4603-8212-45cd0a2d2413 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.365086] env[62385]: DEBUG oslo_vmware.api [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for the task: (returnval){ [ 746.365086] env[62385]: value = "task-1205573" [ 746.365086] env[62385]: _type = "Task" [ 746.365086] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.374625] env[62385]: DEBUG oslo_vmware.api [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205573, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.666707] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance b4970e1a-2b29-44e9-b79e-cda3c32dddcd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 746.679711] env[62385]: DEBUG oslo_vmware.api [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': task-1205569, 'name': PowerOnVM_Task, 'duration_secs': 1.429278} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.680643] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 746.680847] env[62385]: INFO nova.compute.manager [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Took 9.50 seconds to spawn the instance on the hypervisor. [ 746.681053] env[62385]: DEBUG nova.compute.manager [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 746.681874] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdea0b85-7083-425b-9abe-e7a79393fcc5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.737509] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205572, 'name': CreateVM_Task, 'duration_secs': 0.45657} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.738017] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 746.738441] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.738605] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.738942] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 746.739222] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d1b5a1c-889a-4067-93e9-1d20a118cef0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.744823] env[62385]: DEBUG oslo_vmware.api [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 746.744823] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]527fd01c-6b5b-a59c-9e3d-03fec4c5d1df" [ 746.744823] env[62385]: _type = "Task" [ 746.744823] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.754262] env[62385]: DEBUG oslo_vmware.api [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]527fd01c-6b5b-a59c-9e3d-03fec4c5d1df, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.791654] env[62385]: DEBUG nova.network.neutron [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 746.876856] env[62385]: DEBUG oslo_vmware.api [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205573, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.996048] env[62385]: DEBUG nova.network.neutron [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Updating instance_info_cache with network_info: [{"id": "185ea4de-3b36-467d-a112-f439b722671d", "address": "fa:16:3e:0c:87:c8", "network": {"id": "8b2d7a6e-25d4-4a54-ab54-5972612046da", "bridge": "br-int", "label": "tempest-ServersTestJSON-1142134211-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bf8e3c942d3445919cfbe988cca84e90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap185ea4de-3b", "ovs_interfaceid": "185ea4de-3b36-467d-a112-f439b722671d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.171882] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance fd7a77e2-6d76-4d91-bdef-e30333247aa9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 747.202798] env[62385]: INFO nova.compute.manager [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Took 27.95 seconds to build instance. [ 747.257384] env[62385]: DEBUG oslo_vmware.api [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]527fd01c-6b5b-a59c-9e3d-03fec4c5d1df, 'name': SearchDatastore_Task, 'duration_secs': 0.01058} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.257699] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.257962] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 747.258290] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.258548] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.258761] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 747.259076] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4a346eff-907c-4543-a61b-c5bc65408221 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.268572] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 747.268974] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 747.270474] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0110ea20-9842-47f9-959e-121aaa96a88b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.276771] env[62385]: DEBUG oslo_vmware.api [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 747.276771] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5268e20f-0a8d-c91d-cf59-fd06c5925d74" [ 747.276771] env[62385]: _type = "Task" [ 747.276771] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.285139] env[62385]: DEBUG oslo_vmware.api [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5268e20f-0a8d-c91d-cf59-fd06c5925d74, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.376168] env[62385]: DEBUG oslo_vmware.api [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205573, 'name': PowerOnVM_Task, 'duration_secs': 0.62047} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.376476] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 747.376726] env[62385]: INFO nova.compute.manager [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Took 7.59 seconds to spawn the instance on the hypervisor. [ 747.376837] env[62385]: DEBUG nova.compute.manager [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 747.377945] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a278c2cf-9bbf-489b-bedb-a5f5b12c2b00 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.501016] env[62385]: DEBUG oslo_concurrency.lockutils [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Releasing lock "refresh_cache-42ecf594-dff7-4af3-ac56-5bbc45b7a192" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.501016] env[62385]: DEBUG nova.compute.manager [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Instance network_info: |[{"id": "185ea4de-3b36-467d-a112-f439b722671d", "address": "fa:16:3e:0c:87:c8", "network": {"id": "8b2d7a6e-25d4-4a54-ab54-5972612046da", "bridge": "br-int", "label": "tempest-ServersTestJSON-1142134211-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bf8e3c942d3445919cfbe988cca84e90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap185ea4de-3b", "ovs_interfaceid": "185ea4de-3b36-467d-a112-f439b722671d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 747.501263] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0c:87:c8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '185ea4de-3b36-467d-a112-f439b722671d', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 747.512309] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Creating folder: Project (bf8e3c942d3445919cfbe988cca84e90). Parent ref: group-v261107. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 747.512309] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-709feea2-dc42-4ca5-8c6c-42da0ecfe21b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.524555] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Created folder: Project (bf8e3c942d3445919cfbe988cca84e90) in parent group-v261107. [ 747.524872] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Creating folder: Instances. Parent ref: group-v261146. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 747.526486] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c6fe7745-fc31-4aa6-9122-45b991bf7854 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.537163] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Created folder: Instances in parent group-v261146. [ 747.537684] env[62385]: DEBUG oslo.service.loopingcall [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 747.537684] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 747.537919] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4078456b-b8d6-4e0f-983c-f98f53b87e47 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.558879] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 747.558879] env[62385]: value = "task-1205576" [ 747.558879] env[62385]: _type = "Task" [ 747.558879] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.567202] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205576, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.675750] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 747.708138] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7d32c545-4582-42d3-8f48-27261df2ac54 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Lock "749e0976-2e2b-4764-865c-2e630f2edbd1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 151.532s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.791143] env[62385]: DEBUG oslo_vmware.api [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5268e20f-0a8d-c91d-cf59-fd06c5925d74, 'name': SearchDatastore_Task, 'duration_secs': 0.012082} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.791143] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c3300dc8-e48e-49d0-8c49-f834bda778f5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.798091] env[62385]: DEBUG oslo_vmware.api [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 747.798091] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]520352e1-ac45-0111-1051-80f5d77aaf8c" [ 747.798091] env[62385]: _type = "Task" [ 747.798091] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.812839] env[62385]: DEBUG oslo_vmware.api [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]520352e1-ac45-0111-1051-80f5d77aaf8c, 'name': SearchDatastore_Task, 'duration_secs': 0.01287} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.813205] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.813560] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] caca5304-0ad7-427c-812e-de925de63f2c/caca5304-0ad7-427c-812e-de925de63f2c.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 747.813909] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c91e48db-cd77-4d87-a3f2-0230a6183772 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.822753] env[62385]: DEBUG oslo_vmware.api [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 747.822753] env[62385]: value = "task-1205577" [ 747.822753] env[62385]: _type = "Task" [ 747.822753] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.832382] env[62385]: DEBUG oslo_vmware.api [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205577, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.895027] env[62385]: INFO nova.compute.manager [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Took 25.17 seconds to build instance. [ 748.072444] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205576, 'name': CreateVM_Task, 'duration_secs': 0.344847} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.072640] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 748.073530] env[62385]: DEBUG oslo_concurrency.lockutils [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.073783] env[62385]: DEBUG oslo_concurrency.lockutils [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.074191] env[62385]: DEBUG oslo_concurrency.lockutils [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 748.074525] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2c8d135-1007-48db-8dbc-5d0317b2b54b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.080616] env[62385]: DEBUG oslo_vmware.api [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 748.080616] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ad4e43-51dc-06da-66e4-f988831d9f34" [ 748.080616] env[62385]: _type = "Task" [ 748.080616] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.090294] env[62385]: DEBUG oslo_vmware.api [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ad4e43-51dc-06da-66e4-f988831d9f34, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.122980] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9e6a2dab-7bf8-4fe5-a0e1-11e4d30c5c85 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Acquiring lock "e97ff020-61f3-4947-bb82-5c039ad17747" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.123299] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9e6a2dab-7bf8-4fe5-a0e1-11e4d30c5c85 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Lock "e97ff020-61f3-4947-bb82-5c039ad17747" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 748.123503] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9e6a2dab-7bf8-4fe5-a0e1-11e4d30c5c85 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Acquiring lock "e97ff020-61f3-4947-bb82-5c039ad17747-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.123685] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9e6a2dab-7bf8-4fe5-a0e1-11e4d30c5c85 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Lock "e97ff020-61f3-4947-bb82-5c039ad17747-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 748.123855] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9e6a2dab-7bf8-4fe5-a0e1-11e4d30c5c85 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Lock "e97ff020-61f3-4947-bb82-5c039ad17747-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.126197] env[62385]: INFO nova.compute.manager [None req-9e6a2dab-7bf8-4fe5-a0e1-11e4d30c5c85 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Terminating instance [ 748.128242] env[62385]: DEBUG nova.compute.manager [None req-9e6a2dab-7bf8-4fe5-a0e1-11e4d30c5c85 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 748.128439] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-9e6a2dab-7bf8-4fe5-a0e1-11e4d30c5c85 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 748.129314] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17d971c6-bd1e-4112-9d05-f22fc3d78834 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.137477] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e6a2dab-7bf8-4fe5-a0e1-11e4d30c5c85 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 748.137726] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-125ab78b-c510-47d3-88a3-70374c7b7af6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.147084] env[62385]: DEBUG oslo_vmware.api [None req-9e6a2dab-7bf8-4fe5-a0e1-11e4d30c5c85 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Waiting for the task: (returnval){ [ 748.147084] env[62385]: value = "task-1205578" [ 748.147084] env[62385]: _type = "Task" [ 748.147084] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.158208] env[62385]: DEBUG oslo_vmware.api [None req-9e6a2dab-7bf8-4fe5-a0e1-11e4d30c5c85 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': task-1205578, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.178068] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 9d6f098a-0b05-43ef-96b0-9eb99ad3538c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 748.210976] env[62385]: DEBUG nova.compute.manager [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 748.235154] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d7c0ede1-27a5-42bd-99d1-019b1c52ecae tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Acquiring lock "749e0976-2e2b-4764-865c-2e630f2edbd1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.235427] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d7c0ede1-27a5-42bd-99d1-019b1c52ecae tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Lock "749e0976-2e2b-4764-865c-2e630f2edbd1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 748.236141] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d7c0ede1-27a5-42bd-99d1-019b1c52ecae tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Acquiring lock "749e0976-2e2b-4764-865c-2e630f2edbd1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.236415] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d7c0ede1-27a5-42bd-99d1-019b1c52ecae tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Lock "749e0976-2e2b-4764-865c-2e630f2edbd1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 748.236546] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d7c0ede1-27a5-42bd-99d1-019b1c52ecae tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Lock "749e0976-2e2b-4764-865c-2e630f2edbd1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.239091] env[62385]: INFO nova.compute.manager [None req-d7c0ede1-27a5-42bd-99d1-019b1c52ecae tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Terminating instance [ 748.241078] env[62385]: DEBUG nova.compute.manager [None req-d7c0ede1-27a5-42bd-99d1-019b1c52ecae tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 748.241281] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d7c0ede1-27a5-42bd-99d1-019b1c52ecae tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 748.242162] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a492f959-8bcd-4c2e-a4ec-0bee6ac415c1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.252023] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7c0ede1-27a5-42bd-99d1-019b1c52ecae tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 748.252301] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3b973b90-1d25-4e7f-8d86-03f6ff6e7881 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.263458] env[62385]: DEBUG oslo_vmware.api [None req-d7c0ede1-27a5-42bd-99d1-019b1c52ecae tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Waiting for the task: (returnval){ [ 748.263458] env[62385]: value = "task-1205579" [ 748.263458] env[62385]: _type = "Task" [ 748.263458] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.277441] env[62385]: DEBUG oslo_vmware.api [None req-d7c0ede1-27a5-42bd-99d1-019b1c52ecae tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': task-1205579, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.337778] env[62385]: DEBUG oslo_vmware.api [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205577, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.400181] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67c0fcd4-749e-4176-9f4a-8e6387e43eac tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Lock "92057af7-28a3-4643-9cda-d3d868d0cfc9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 145.171s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.593329] env[62385]: DEBUG oslo_vmware.api [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ad4e43-51dc-06da-66e4-f988831d9f34, 'name': SearchDatastore_Task, 'duration_secs': 0.064781} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.593641] env[62385]: DEBUG oslo_concurrency.lockutils [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.593873] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 748.594120] env[62385]: DEBUG oslo_concurrency.lockutils [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.594269] env[62385]: DEBUG oslo_concurrency.lockutils [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.594473] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 748.594708] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d71f90e5-142a-480a-b230-4f4430971f87 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.604182] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 748.604182] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 748.604605] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-59633b5d-74f3-4f73-bb71-df7c0ac4ab4c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.610791] env[62385]: DEBUG oslo_vmware.api [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 748.610791] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52878422-e002-f716-c174-90c291424bd1" [ 748.610791] env[62385]: _type = "Task" [ 748.610791] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.618689] env[62385]: DEBUG oslo_vmware.api [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52878422-e002-f716-c174-90c291424bd1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.660875] env[62385]: DEBUG oslo_vmware.api [None req-9e6a2dab-7bf8-4fe5-a0e1-11e4d30c5c85 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': task-1205578, 'name': PowerOffVM_Task, 'duration_secs': 0.3651} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.661177] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e6a2dab-7bf8-4fe5-a0e1-11e4d30c5c85 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 748.661414] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-9e6a2dab-7bf8-4fe5-a0e1-11e4d30c5c85 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 748.661599] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3b7efbb1-fd8d-4542-9ff3-e55e1e2be2a2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.681560] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 2b744879-01d2-440e-8adf-58438b922888 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 748.729260] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-9e6a2dab-7bf8-4fe5-a0e1-11e4d30c5c85 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 748.729260] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-9e6a2dab-7bf8-4fe5-a0e1-11e4d30c5c85 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 748.729431] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e6a2dab-7bf8-4fe5-a0e1-11e4d30c5c85 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Deleting the datastore file [datastore1] e97ff020-61f3-4947-bb82-5c039ad17747 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 748.729713] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-698317f4-da76-4913-be91-f5d506c74bf7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.733940] env[62385]: DEBUG oslo_concurrency.lockutils [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.738457] env[62385]: DEBUG oslo_vmware.api [None req-9e6a2dab-7bf8-4fe5-a0e1-11e4d30c5c85 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Waiting for the task: (returnval){ [ 748.738457] env[62385]: value = "task-1205581" [ 748.738457] env[62385]: _type = "Task" [ 748.738457] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.746657] env[62385]: DEBUG oslo_vmware.api [None req-9e6a2dab-7bf8-4fe5-a0e1-11e4d30c5c85 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': task-1205581, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.774423] env[62385]: DEBUG oslo_vmware.api [None req-d7c0ede1-27a5-42bd-99d1-019b1c52ecae tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': task-1205579, 'name': PowerOffVM_Task, 'duration_secs': 0.258444} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.774685] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7c0ede1-27a5-42bd-99d1-019b1c52ecae tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 748.774848] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d7c0ede1-27a5-42bd-99d1-019b1c52ecae tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 748.775108] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-524e2ed4-08c4-4fd9-97d7-0694b6e48fa0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.835025] env[62385]: DEBUG oslo_vmware.api [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205577, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.545421} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.835025] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] caca5304-0ad7-427c-812e-de925de63f2c/caca5304-0ad7-427c-812e-de925de63f2c.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 748.835210] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 748.835349] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0a8cd3dd-f663-4a1f-b753-fdc7a3bb083e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.843529] env[62385]: DEBUG oslo_vmware.api [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 748.843529] env[62385]: value = "task-1205583" [ 748.843529] env[62385]: _type = "Task" [ 748.843529] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.852464] env[62385]: DEBUG oslo_vmware.api [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205583, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.890719] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d7c0ede1-27a5-42bd-99d1-019b1c52ecae tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 748.890898] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d7c0ede1-27a5-42bd-99d1-019b1c52ecae tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 748.891858] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7c0ede1-27a5-42bd-99d1-019b1c52ecae tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Deleting the datastore file [datastore1] 749e0976-2e2b-4764-865c-2e630f2edbd1 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 748.891858] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-650c8eed-65f6-4332-838f-3d1ef78e96fb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.898749] env[62385]: DEBUG oslo_vmware.api [None req-d7c0ede1-27a5-42bd-99d1-019b1c52ecae tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Waiting for the task: (returnval){ [ 748.898749] env[62385]: value = "task-1205584" [ 748.898749] env[62385]: _type = "Task" [ 748.898749] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.902062] env[62385]: DEBUG nova.compute.manager [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 748.910046] env[62385]: DEBUG oslo_vmware.api [None req-d7c0ede1-27a5-42bd-99d1-019b1c52ecae tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': task-1205584, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.128462] env[62385]: DEBUG oslo_vmware.api [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52878422-e002-f716-c174-90c291424bd1, 'name': SearchDatastore_Task, 'duration_secs': 0.018194} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.129865] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4631f6fc-c750-43dc-ba97-1f56c6f0875c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.139061] env[62385]: DEBUG oslo_vmware.api [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 749.139061] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]522f7cf5-5ab3-117f-7072-ebeb3de788c0" [ 749.139061] env[62385]: _type = "Task" [ 749.139061] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.147429] env[62385]: DEBUG oslo_vmware.api [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]522f7cf5-5ab3-117f-7072-ebeb3de788c0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.184486] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance e4089e71-eea1-4e16-b90c-97966333f2bc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 749.250864] env[62385]: DEBUG oslo_vmware.api [None req-9e6a2dab-7bf8-4fe5-a0e1-11e4d30c5c85 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': task-1205581, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.177345} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.251104] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e6a2dab-7bf8-4fe5-a0e1-11e4d30c5c85 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 749.251272] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-9e6a2dab-7bf8-4fe5-a0e1-11e4d30c5c85 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 749.251496] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-9e6a2dab-7bf8-4fe5-a0e1-11e4d30c5c85 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 749.252393] env[62385]: INFO nova.compute.manager [None req-9e6a2dab-7bf8-4fe5-a0e1-11e4d30c5c85 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Took 1.12 seconds to destroy the instance on the hypervisor. [ 749.252393] env[62385]: DEBUG oslo.service.loopingcall [None req-9e6a2dab-7bf8-4fe5-a0e1-11e4d30c5c85 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 749.252566] env[62385]: DEBUG nova.compute.manager [-] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 749.252672] env[62385]: DEBUG nova.network.neutron [-] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 749.354477] env[62385]: DEBUG oslo_vmware.api [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205583, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.093157} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.354744] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 749.355566] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6176366-81be-4de5-94f4-8fbc5fb12d9d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.382764] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Reconfiguring VM instance instance-00000030 to attach disk [datastore1] caca5304-0ad7-427c-812e-de925de63f2c/caca5304-0ad7-427c-812e-de925de63f2c.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 749.383127] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a455eb07-6c45-43fa-ba5e-f0f6cc2ca301 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.411767] env[62385]: DEBUG oslo_vmware.api [None req-d7c0ede1-27a5-42bd-99d1-019b1c52ecae tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Task: {'id': task-1205584, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.200593} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.412238] env[62385]: DEBUG oslo_vmware.api [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 749.412238] env[62385]: value = "task-1205585" [ 749.412238] env[62385]: _type = "Task" [ 749.412238] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.414844] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7c0ede1-27a5-42bd-99d1-019b1c52ecae tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 749.415065] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d7c0ede1-27a5-42bd-99d1-019b1c52ecae tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 749.415256] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d7c0ede1-27a5-42bd-99d1-019b1c52ecae tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 749.415428] env[62385]: INFO nova.compute.manager [None req-d7c0ede1-27a5-42bd-99d1-019b1c52ecae tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Took 1.17 seconds to destroy the instance on the hypervisor. [ 749.415664] env[62385]: DEBUG oslo.service.loopingcall [None req-d7c0ede1-27a5-42bd-99d1-019b1c52ecae tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 749.416620] env[62385]: DEBUG nova.compute.manager [-] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 749.416730] env[62385]: DEBUG nova.network.neutron [-] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 749.428452] env[62385]: DEBUG oslo_vmware.api [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205585, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.432496] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 749.658298] env[62385]: DEBUG oslo_vmware.api [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]522f7cf5-5ab3-117f-7072-ebeb3de788c0, 'name': SearchDatastore_Task, 'duration_secs': 0.010171} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.659584] env[62385]: DEBUG oslo_concurrency.lockutils [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 749.659584] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 42ecf594-dff7-4af3-ac56-5bbc45b7a192/42ecf594-dff7-4af3-ac56-5bbc45b7a192.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 749.659584] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4014dfc8-d5b8-4d71-a9b8-09eade52cc1a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.668514] env[62385]: DEBUG oslo_vmware.api [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 749.668514] env[62385]: value = "task-1205586" [ 749.668514] env[62385]: _type = "Task" [ 749.668514] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.677603] env[62385]: DEBUG oslo_vmware.api [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1205586, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.688161] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance f2be792c-a723-4a69-b255-bbe7f0e8692d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 749.778606] env[62385]: DEBUG nova.compute.manager [req-a86321c9-0dca-4f8f-9c30-c31f38cf636c req-0d639f1a-c928-422f-b17e-63aed856c029 service nova] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Received event network-vif-deleted-5a29f755-1606-46ef-9967-30420d891d4f {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 749.778784] env[62385]: INFO nova.compute.manager [req-a86321c9-0dca-4f8f-9c30-c31f38cf636c req-0d639f1a-c928-422f-b17e-63aed856c029 service nova] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Neutron deleted interface 5a29f755-1606-46ef-9967-30420d891d4f; detaching it from the instance and deleting it from the info cache [ 749.779571] env[62385]: DEBUG nova.network.neutron [req-a86321c9-0dca-4f8f-9c30-c31f38cf636c req-0d639f1a-c928-422f-b17e-63aed856c029 service nova] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.838644] env[62385]: DEBUG nova.compute.manager [req-4f80ff31-bfdb-4eca-b682-f4ad7cdfbc0d req-3d4850a2-c879-4a8e-8004-8fcfe5692616 service nova] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Received event network-vif-deleted-6dd0572e-2124-47c7-b8e5-900aa035f6c4 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 749.839181] env[62385]: INFO nova.compute.manager [req-4f80ff31-bfdb-4eca-b682-f4ad7cdfbc0d req-3d4850a2-c879-4a8e-8004-8fcfe5692616 service nova] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Neutron deleted interface 6dd0572e-2124-47c7-b8e5-900aa035f6c4; detaching it from the instance and deleting it from the info cache [ 749.841683] env[62385]: DEBUG nova.network.neutron [req-4f80ff31-bfdb-4eca-b682-f4ad7cdfbc0d req-3d4850a2-c879-4a8e-8004-8fcfe5692616 service nova] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.930281] env[62385]: DEBUG oslo_vmware.api [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205585, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.181174] env[62385]: DEBUG nova.network.neutron [-] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.186757] env[62385]: DEBUG oslo_vmware.api [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1205586, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.191371] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 55d62bba-ff30-42bf-b1cb-7567988a9361 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 750.193155] env[62385]: DEBUG nova.network.neutron [-] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.282712] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9b2b4891-f97b-4d61-8f86-3d7cb0a94729 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.292979] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4256cbef-1428-401e-b531-5a82f78ae944 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.321954] env[62385]: DEBUG nova.compute.manager [req-a86321c9-0dca-4f8f-9c30-c31f38cf636c req-0d639f1a-c928-422f-b17e-63aed856c029 service nova] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Detach interface failed, port_id=5a29f755-1606-46ef-9967-30420d891d4f, reason: Instance 749e0976-2e2b-4764-865c-2e630f2edbd1 could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 750.343451] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1fce2d38-ddf8-46b8-bdc6-ee41246e6447 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.353367] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd72448a-ff22-4c4c-ad86-4f9226a0ea6e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.383241] env[62385]: DEBUG nova.compute.manager [req-4f80ff31-bfdb-4eca-b682-f4ad7cdfbc0d req-3d4850a2-c879-4a8e-8004-8fcfe5692616 service nova] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Detach interface failed, port_id=6dd0572e-2124-47c7-b8e5-900aa035f6c4, reason: Instance e97ff020-61f3-4947-bb82-5c039ad17747 could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 750.426695] env[62385]: DEBUG oslo_vmware.api [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205585, 'name': ReconfigVM_Task, 'duration_secs': 0.629237} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.429026] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Reconfigured VM instance instance-00000030 to attach disk [datastore1] caca5304-0ad7-427c-812e-de925de63f2c/caca5304-0ad7-427c-812e-de925de63f2c.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 750.429026] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-700740f7-0b41-430a-ac05-1b3a886f8c60 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.435349] env[62385]: DEBUG oslo_vmware.api [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 750.435349] env[62385]: value = "task-1205587" [ 750.435349] env[62385]: _type = "Task" [ 750.435349] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.443896] env[62385]: DEBUG oslo_vmware.api [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205587, 'name': Rename_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.682569] env[62385]: DEBUG oslo_vmware.api [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1205586, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.544751} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.682845] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 42ecf594-dff7-4af3-ac56-5bbc45b7a192/42ecf594-dff7-4af3-ac56-5bbc45b7a192.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 750.683083] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 750.683442] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7a7dc232-ebb1-4aa3-96f6-3bbfce719ee1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.687534] env[62385]: INFO nova.compute.manager [-] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Took 1.27 seconds to deallocate network for instance. [ 750.693970] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 750.694962] env[62385]: DEBUG oslo_vmware.api [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 750.694962] env[62385]: value = "task-1205588" [ 750.694962] env[62385]: _type = "Task" [ 750.694962] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.695720] env[62385]: INFO nova.compute.manager [-] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Took 1.44 seconds to deallocate network for instance. [ 750.725909] env[62385]: DEBUG oslo_vmware.api [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1205588, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.945307] env[62385]: DEBUG oslo_vmware.api [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205587, 'name': Rename_Task, 'duration_secs': 0.352336} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.945645] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 750.945807] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4501e7f3-bdea-44b6-a3e6-38f6df061bd8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.953017] env[62385]: DEBUG oslo_vmware.api [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 750.953017] env[62385]: value = "task-1205589" [ 750.953017] env[62385]: _type = "Task" [ 750.953017] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.961808] env[62385]: DEBUG oslo_vmware.api [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205589, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.196709] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d7c0ede1-27a5-42bd-99d1-019b1c52ecae tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.197501] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 679150f5-ad17-45e6-9e3d-9f6e28b1aee3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 751.197732] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Total usable vcpus: 48, total allocated vcpus: 12 {{(pid=62385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 751.197961] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2880MB phys_disk=200GB used_disk=12GB total_vcpus=48 used_vcpus=12 pci_stats=[] {{(pid=62385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 751.207472] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9e6a2dab-7bf8-4fe5-a0e1-11e4d30c5c85 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.211337] env[62385]: DEBUG oslo_vmware.api [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1205588, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.270512} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.211608] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 751.212437] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1297f5c5-1d2c-4c1e-98da-8376531afd77 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.237429] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Reconfiguring VM instance instance-00000034 to attach disk [datastore1] 42ecf594-dff7-4af3-ac56-5bbc45b7a192/42ecf594-dff7-4af3-ac56-5bbc45b7a192.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 751.240510] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bd572934-c3dd-4609-94bf-5b429e02fd29 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.262040] env[62385]: DEBUG oslo_vmware.api [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 751.262040] env[62385]: value = "task-1205590" [ 751.262040] env[62385]: _type = "Task" [ 751.262040] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.273408] env[62385]: DEBUG oslo_vmware.api [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1205590, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.462719] env[62385]: DEBUG oslo_vmware.api [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205589, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.560828] env[62385]: DEBUG nova.compute.manager [None req-470c003f-95c5-4253-ac89-1328d5d25cf1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 751.561844] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-732cf415-fa48-4572-b776-afd64363c0dd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.613720] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-904849e3-0bcf-47bd-9676-b82534cfed71 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.623462] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2319e9f-6b70-4553-90b6-157a0246c1aa {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.657482] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cce0ae7d-09a8-4f63-825b-6dae03876b99 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.666110] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91e587a7-4c7e-43d5-867d-7891b92af352 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.681445] env[62385]: DEBUG nova.compute.provider_tree [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Updating inventory in ProviderTree for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 751.777281] env[62385]: DEBUG oslo_vmware.api [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1205590, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.965518] env[62385]: DEBUG oslo_vmware.api [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205589, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.075645] env[62385]: INFO nova.compute.manager [None req-470c003f-95c5-4253-ac89-1328d5d25cf1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] instance snapshotting [ 752.078692] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74ee4af4-cdca-4b79-8bf5-1d951ceb7c54 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.098749] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4adacd1b-1adc-4fe2-8b5a-a7dcb3626a78 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.213713] env[62385]: DEBUG nova.scheduler.client.report [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Updated inventory for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a with generation 79 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 752.213974] env[62385]: DEBUG nova.compute.provider_tree [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Updating resource provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a generation from 79 to 80 during operation: update_inventory {{(pid=62385) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 752.214162] env[62385]: DEBUG nova.compute.provider_tree [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Updating inventory in ProviderTree for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 752.274517] env[62385]: DEBUG oslo_vmware.api [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1205590, 'name': ReconfigVM_Task, 'duration_secs': 0.751497} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.274807] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Reconfigured VM instance instance-00000034 to attach disk [datastore1] 42ecf594-dff7-4af3-ac56-5bbc45b7a192/42ecf594-dff7-4af3-ac56-5bbc45b7a192.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 752.275463] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-78714667-d2b4-4bee-94e2-27f96fc81a1f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.282874] env[62385]: DEBUG oslo_vmware.api [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 752.282874] env[62385]: value = "task-1205591" [ 752.282874] env[62385]: _type = "Task" [ 752.282874] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.290927] env[62385]: DEBUG oslo_vmware.api [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1205591, 'name': Rename_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.464938] env[62385]: DEBUG oslo_vmware.api [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205589, 'name': PowerOnVM_Task, 'duration_secs': 1.068957} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.465710] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 752.466084] env[62385]: DEBUG nova.compute.manager [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 752.466851] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfcbe875-bbed-496d-b790-bf2d9dd23bff {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.610029] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-470c003f-95c5-4253-ac89-1328d5d25cf1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Creating Snapshot of the VM instance {{(pid=62385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 752.610029] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-6f081261-416a-43f3-b772-d839f0fe2c68 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.620259] env[62385]: DEBUG oslo_vmware.api [None req-470c003f-95c5-4253-ac89-1328d5d25cf1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for the task: (returnval){ [ 752.620259] env[62385]: value = "task-1205592" [ 752.620259] env[62385]: _type = "Task" [ 752.620259] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.630246] env[62385]: DEBUG oslo_vmware.api [None req-470c003f-95c5-4253-ac89-1328d5d25cf1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205592, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.719947] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62385) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 752.720323] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 11.163s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.720664] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.338s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.722425] env[62385]: INFO nova.compute.claims [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 752.795498] env[62385]: DEBUG oslo_vmware.api [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1205591, 'name': Rename_Task, 'duration_secs': 0.169788} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.795842] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 752.796145] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a6e429ed-dfd7-4a29-8bdd-3f87963a3cb0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.806087] env[62385]: DEBUG oslo_vmware.api [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 752.806087] env[62385]: value = "task-1205593" [ 752.806087] env[62385]: _type = "Task" [ 752.806087] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.814445] env[62385]: DEBUG oslo_vmware.api [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1205593, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.984322] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.131397] env[62385]: DEBUG oslo_vmware.api [None req-470c003f-95c5-4253-ac89-1328d5d25cf1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205592, 'name': CreateSnapshot_Task, 'duration_secs': 0.483114} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.131397] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-470c003f-95c5-4253-ac89-1328d5d25cf1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Created Snapshot of the VM instance {{(pid=62385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 753.131969] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34d6baf8-bc9b-4023-b9ec-c002d87f5354 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.323911] env[62385]: DEBUG oslo_vmware.api [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1205593, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.650661] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-470c003f-95c5-4253-ac89-1328d5d25cf1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Creating linked-clone VM from snapshot {{(pid=62385) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 753.650979] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-9d6af734-74f1-4b4c-8730-d83d84842c44 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.661483] env[62385]: DEBUG oslo_vmware.api [None req-470c003f-95c5-4253-ac89-1328d5d25cf1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for the task: (returnval){ [ 753.661483] env[62385]: value = "task-1205594" [ 753.661483] env[62385]: _type = "Task" [ 753.661483] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.669917] env[62385]: DEBUG oslo_vmware.api [None req-470c003f-95c5-4253-ac89-1328d5d25cf1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205594, 'name': CloneVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.749707] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a40d5bae-eeab-45c3-9849-2e493ad05ea0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "caca5304-0ad7-427c-812e-de925de63f2c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.749989] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a40d5bae-eeab-45c3-9849-2e493ad05ea0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "caca5304-0ad7-427c-812e-de925de63f2c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.750257] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a40d5bae-eeab-45c3-9849-2e493ad05ea0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "caca5304-0ad7-427c-812e-de925de63f2c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.750454] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a40d5bae-eeab-45c3-9849-2e493ad05ea0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "caca5304-0ad7-427c-812e-de925de63f2c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.750628] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a40d5bae-eeab-45c3-9849-2e493ad05ea0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "caca5304-0ad7-427c-812e-de925de63f2c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.753406] env[62385]: INFO nova.compute.manager [None req-a40d5bae-eeab-45c3-9849-2e493ad05ea0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Terminating instance [ 753.755306] env[62385]: DEBUG nova.compute.manager [None req-a40d5bae-eeab-45c3-9849-2e493ad05ea0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 753.755476] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a40d5bae-eeab-45c3-9849-2e493ad05ea0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 753.756397] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc7259a4-7d66-4bd1-9d99-3b8c34a9d580 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.767388] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a40d5bae-eeab-45c3-9849-2e493ad05ea0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 753.767635] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f68d97b1-ca5f-4a74-9946-76f863fa6ee9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.775620] env[62385]: DEBUG oslo_vmware.api [None req-a40d5bae-eeab-45c3-9849-2e493ad05ea0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 753.775620] env[62385]: value = "task-1205595" [ 753.775620] env[62385]: _type = "Task" [ 753.775620] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.785283] env[62385]: DEBUG oslo_vmware.api [None req-a40d5bae-eeab-45c3-9849-2e493ad05ea0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205595, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.821167] env[62385]: DEBUG oslo_vmware.api [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1205593, 'name': PowerOnVM_Task, 'duration_secs': 0.523568} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.824427] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 753.824536] env[62385]: INFO nova.compute.manager [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Took 10.23 seconds to spawn the instance on the hypervisor. [ 753.824741] env[62385]: DEBUG nova.compute.manager [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 753.826497] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6471649c-7ca9-4624-9aa8-70eee5717478 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.168063] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0275863-580c-4e2d-ab3d-02de34e8128a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.179900] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58cf198c-d821-40ae-89df-f7ec652caed0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.183224] env[62385]: DEBUG oslo_vmware.api [None req-470c003f-95c5-4253-ac89-1328d5d25cf1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205594, 'name': CloneVM_Task} progress is 94%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.210615] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa160a3f-cbbc-4bb3-a929-90a1958d3e63 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.219630] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9495fb41-be15-415a-b7e0-459247cb95f4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.232639] env[62385]: DEBUG nova.compute.provider_tree [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 754.285794] env[62385]: DEBUG oslo_vmware.api [None req-a40d5bae-eeab-45c3-9849-2e493ad05ea0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205595, 'name': PowerOffVM_Task, 'duration_secs': 0.3694} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.286024] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a40d5bae-eeab-45c3-9849-2e493ad05ea0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 754.286177] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a40d5bae-eeab-45c3-9849-2e493ad05ea0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 754.286450] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0fba85ee-6527-478e-9d05-e321deb80a6a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.347313] env[62385]: INFO nova.compute.manager [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Took 30.91 seconds to build instance. [ 754.362669] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a40d5bae-eeab-45c3-9849-2e493ad05ea0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 754.362997] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a40d5bae-eeab-45c3-9849-2e493ad05ea0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 754.363293] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a40d5bae-eeab-45c3-9849-2e493ad05ea0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Deleting the datastore file [datastore1] caca5304-0ad7-427c-812e-de925de63f2c {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 754.364562] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1d2bc212-0274-418a-88fb-e1d11c1813f5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.374166] env[62385]: DEBUG oslo_vmware.api [None req-a40d5bae-eeab-45c3-9849-2e493ad05ea0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 754.374166] env[62385]: value = "task-1205597" [ 754.374166] env[62385]: _type = "Task" [ 754.374166] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.386439] env[62385]: DEBUG oslo_vmware.api [None req-a40d5bae-eeab-45c3-9849-2e493ad05ea0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205597, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.480538] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "5a82fc98-e656-4476-b7dc-99466e9d6afd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.480788] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "5a82fc98-e656-4476-b7dc-99466e9d6afd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.672928] env[62385]: DEBUG oslo_vmware.api [None req-470c003f-95c5-4253-ac89-1328d5d25cf1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205594, 'name': CloneVM_Task, 'duration_secs': 0.913565} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.673218] env[62385]: INFO nova.virt.vmwareapi.vmops [None req-470c003f-95c5-4253-ac89-1328d5d25cf1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Created linked-clone VM from snapshot [ 754.673988] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-637ce6cd-a4b2-49c8-bfb3-c614811d3071 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.681681] env[62385]: DEBUG nova.virt.vmwareapi.images [None req-470c003f-95c5-4253-ac89-1328d5d25cf1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Uploading image 396dbba4-ec88-4bae-99a7-4cfbaf922102 {{(pid=62385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 754.695829] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-470c003f-95c5-4253-ac89-1328d5d25cf1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Destroying the VM {{(pid=62385) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 754.696120] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-dac910e7-cbe1-440f-b490-cd2d5356f237 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.718044] env[62385]: DEBUG oslo_vmware.api [None req-470c003f-95c5-4253-ac89-1328d5d25cf1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for the task: (returnval){ [ 754.718044] env[62385]: value = "task-1205598" [ 754.718044] env[62385]: _type = "Task" [ 754.718044] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.726363] env[62385]: DEBUG oslo_vmware.api [None req-470c003f-95c5-4253-ac89-1328d5d25cf1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205598, 'name': Destroy_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.735337] env[62385]: DEBUG nova.scheduler.client.report [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 754.849218] env[62385]: DEBUG oslo_concurrency.lockutils [None req-57427494-e9cb-4967-ada5-6c98974b1cc7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "42ecf594-dff7-4af3-ac56-5bbc45b7a192" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 129.285s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.884711] env[62385]: DEBUG oslo_vmware.api [None req-a40d5bae-eeab-45c3-9849-2e493ad05ea0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205597, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13912} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.884952] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a40d5bae-eeab-45c3-9849-2e493ad05ea0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 754.885170] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a40d5bae-eeab-45c3-9849-2e493ad05ea0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 754.885349] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a40d5bae-eeab-45c3-9849-2e493ad05ea0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 754.885518] env[62385]: INFO nova.compute.manager [None req-a40d5bae-eeab-45c3-9849-2e493ad05ea0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Took 1.13 seconds to destroy the instance on the hypervisor. [ 754.885777] env[62385]: DEBUG oslo.service.loopingcall [None req-a40d5bae-eeab-45c3-9849-2e493ad05ea0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 754.886040] env[62385]: DEBUG nova.compute.manager [-] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 754.886148] env[62385]: DEBUG nova.network.neutron [-] [instance: caca5304-0ad7-427c-812e-de925de63f2c] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 755.229930] env[62385]: DEBUG oslo_vmware.api [None req-470c003f-95c5-4253-ac89-1328d5d25cf1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205598, 'name': Destroy_Task} progress is 33%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.240992] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.519s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.240992] env[62385]: DEBUG nova.compute.manager [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 755.245125] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.181s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 755.246363] env[62385]: INFO nova.compute.claims [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 755.352289] env[62385]: DEBUG nova.compute.manager [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 755.434987] env[62385]: DEBUG nova.compute.manager [req-68efa32f-8ea4-4251-b778-f54bee7298ec req-80fdae2f-8e66-4092-b82b-b4edabd3f81e service nova] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Received event network-vif-deleted-35f89d27-3056-4630-b7e1-ee1aa28645b8 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 755.435210] env[62385]: INFO nova.compute.manager [req-68efa32f-8ea4-4251-b778-f54bee7298ec req-80fdae2f-8e66-4092-b82b-b4edabd3f81e service nova] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Neutron deleted interface 35f89d27-3056-4630-b7e1-ee1aa28645b8; detaching it from the instance and deleting it from the info cache [ 755.435392] env[62385]: DEBUG nova.network.neutron [req-68efa32f-8ea4-4251-b778-f54bee7298ec req-80fdae2f-8e66-4092-b82b-b4edabd3f81e service nova] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.730214] env[62385]: DEBUG oslo_vmware.api [None req-470c003f-95c5-4253-ac89-1328d5d25cf1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205598, 'name': Destroy_Task, 'duration_secs': 0.826589} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.730506] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-470c003f-95c5-4253-ac89-1328d5d25cf1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Destroyed the VM [ 755.730826] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-470c003f-95c5-4253-ac89-1328d5d25cf1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Deleting Snapshot of the VM instance {{(pid=62385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 755.731093] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-7d5a3435-cde0-43cb-b7e7-fb4ab2b9b4c2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.739245] env[62385]: DEBUG oslo_vmware.api [None req-470c003f-95c5-4253-ac89-1328d5d25cf1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for the task: (returnval){ [ 755.739245] env[62385]: value = "task-1205599" [ 755.739245] env[62385]: _type = "Task" [ 755.739245] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.500270] env[62385]: DEBUG nova.network.neutron [-] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.502345] env[62385]: DEBUG nova.compute.utils [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 756.507750] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c3bb0442-402a-44c3-a5ec-63b1cb851d77 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "42ecf594-dff7-4af3-ac56-5bbc45b7a192" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 756.508008] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c3bb0442-402a-44c3-a5ec-63b1cb851d77 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "42ecf594-dff7-4af3-ac56-5bbc45b7a192" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 756.508195] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c3bb0442-402a-44c3-a5ec-63b1cb851d77 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "42ecf594-dff7-4af3-ac56-5bbc45b7a192-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 756.508375] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c3bb0442-402a-44c3-a5ec-63b1cb851d77 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "42ecf594-dff7-4af3-ac56-5bbc45b7a192-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 756.508538] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c3bb0442-402a-44c3-a5ec-63b1cb851d77 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "42ecf594-dff7-4af3-ac56-5bbc45b7a192-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.512509] env[62385]: DEBUG nova.compute.manager [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 756.512679] env[62385]: DEBUG nova.network.neutron [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 756.514543] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-712e8ff8-d5a2-41d7-a209-bb0c61b62811 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.517309] env[62385]: INFO nova.compute.manager [None req-c3bb0442-402a-44c3-a5ec-63b1cb851d77 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Terminating instance [ 756.522910] env[62385]: DEBUG oslo_vmware.api [None req-470c003f-95c5-4253-ac89-1328d5d25cf1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205599, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.523120] env[62385]: WARNING oslo_vmware.common.loopingcall [None req-470c003f-95c5-4253-ac89-1328d5d25cf1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] task run outlasted interval by 0.28362600000000004 sec [ 756.523671] env[62385]: DEBUG nova.compute.manager [None req-c3bb0442-402a-44c3-a5ec-63b1cb851d77 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 756.523843] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c3bb0442-402a-44c3-a5ec-63b1cb851d77 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 756.525532] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0f3e918-c7ba-4856-8ac8-e8387d77318f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.535267] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f18e3858-e567-49af-9862-68b2aa3796dc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.545651] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 756.552525] env[62385]: DEBUG oslo_vmware.api [None req-470c003f-95c5-4253-ac89-1328d5d25cf1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205599, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.554796] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3bb0442-402a-44c3-a5ec-63b1cb851d77 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 756.554906] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-59cbdabc-dba5-46cc-9949-ee51c7d4e72c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.564168] env[62385]: DEBUG nova.compute.manager [req-68efa32f-8ea4-4251-b778-f54bee7298ec req-80fdae2f-8e66-4092-b82b-b4edabd3f81e service nova] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Detach interface failed, port_id=35f89d27-3056-4630-b7e1-ee1aa28645b8, reason: Instance caca5304-0ad7-427c-812e-de925de63f2c could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 756.570424] env[62385]: DEBUG oslo_vmware.api [None req-c3bb0442-402a-44c3-a5ec-63b1cb851d77 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 756.570424] env[62385]: value = "task-1205600" [ 756.570424] env[62385]: _type = "Task" [ 756.570424] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.578757] env[62385]: DEBUG oslo_vmware.api [None req-c3bb0442-402a-44c3-a5ec-63b1cb851d77 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1205600, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.581096] env[62385]: DEBUG nova.policy [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '77209f14681043328e5269b63fff4740', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6f08231c6ff14924a8e51b77ad84bdb9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 757.001287] env[62385]: DEBUG nova.network.neutron [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Successfully created port: 81b5c1c2-0651-4034-87db-8faf51c3c7d5 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 757.012941] env[62385]: INFO nova.compute.manager [-] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Took 2.13 seconds to deallocate network for instance. [ 757.016118] env[62385]: DEBUG nova.compute.manager [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 757.029754] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 757.029877] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Cleaning up deleted instances {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 757.037672] env[62385]: DEBUG oslo_vmware.api [None req-470c003f-95c5-4253-ac89-1328d5d25cf1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205599, 'name': RemoveSnapshot_Task, 'duration_secs': 1.071164} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.037944] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-470c003f-95c5-4253-ac89-1328d5d25cf1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Deleted Snapshot of the VM instance {{(pid=62385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 757.083520] env[62385]: DEBUG oslo_vmware.api [None req-c3bb0442-402a-44c3-a5ec-63b1cb851d77 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1205600, 'name': PowerOffVM_Task, 'duration_secs': 0.285216} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.087983] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3bb0442-402a-44c3-a5ec-63b1cb851d77 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 757.087983] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c3bb0442-402a-44c3-a5ec-63b1cb851d77 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 757.087983] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d3a1ef6b-9711-465c-b482-810210b466b0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.161126] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c3bb0442-402a-44c3-a5ec-63b1cb851d77 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 757.161292] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c3bb0442-402a-44c3-a5ec-63b1cb851d77 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 757.161437] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3bb0442-402a-44c3-a5ec-63b1cb851d77 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Deleting the datastore file [datastore1] 42ecf594-dff7-4af3-ac56-5bbc45b7a192 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 757.161694] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f0837747-8514-4c75-8e55-eb18f1bc854b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.172116] env[62385]: DEBUG oslo_vmware.api [None req-c3bb0442-402a-44c3-a5ec-63b1cb851d77 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 757.172116] env[62385]: value = "task-1205602" [ 757.172116] env[62385]: _type = "Task" [ 757.172116] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.184161] env[62385]: DEBUG oslo_vmware.api [None req-c3bb0442-402a-44c3-a5ec-63b1cb851d77 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1205602, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.519954] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55defc43-195c-4bf6-b251-98ffd254f3cb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.527763] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a40d5bae-eeab-45c3-9849-2e493ad05ea0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.530816] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c75a04b-9da8-4185-91b9-72b787aaba13 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.535130] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] There are 3 instances to clean {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 757.535279] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: f0f0eed0-a76a-42bc-9cca-91b93d7ad64c] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 757.568420] env[62385]: WARNING nova.compute.manager [None req-470c003f-95c5-4253-ac89-1328d5d25cf1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Image not found during snapshot: nova.exception.ImageNotFound: Image 396dbba4-ec88-4bae-99a7-4cfbaf922102 could not be found. [ 757.570803] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7aeb13a-b614-4477-b7be-c7bff595f45f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.581758] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5f42526-7c7e-450f-8c1c-a16e94fd90d6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.599249] env[62385]: DEBUG nova.compute.provider_tree [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 757.682296] env[62385]: DEBUG oslo_vmware.api [None req-c3bb0442-402a-44c3-a5ec-63b1cb851d77 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1205602, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154118} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.682592] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3bb0442-402a-44c3-a5ec-63b1cb851d77 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 757.682732] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c3bb0442-402a-44c3-a5ec-63b1cb851d77 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 757.682903] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c3bb0442-402a-44c3-a5ec-63b1cb851d77 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 757.683090] env[62385]: INFO nova.compute.manager [None req-c3bb0442-402a-44c3-a5ec-63b1cb851d77 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Took 1.16 seconds to destroy the instance on the hypervisor. [ 757.683333] env[62385]: DEBUG oslo.service.loopingcall [None req-c3bb0442-402a-44c3-a5ec-63b1cb851d77 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 757.683520] env[62385]: DEBUG nova.compute.manager [-] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 757.683612] env[62385]: DEBUG nova.network.neutron [-] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 757.942048] env[62385]: DEBUG nova.compute.manager [req-d8e1a38e-7f01-452e-bb0a-aefb8fceb303 req-3dd52a16-4910-4544-9c4e-0bd01ab792dd service nova] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Received event network-vif-deleted-185ea4de-3b36-467d-a112-f439b722671d {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 757.942243] env[62385]: INFO nova.compute.manager [req-d8e1a38e-7f01-452e-bb0a-aefb8fceb303 req-3dd52a16-4910-4544-9c4e-0bd01ab792dd service nova] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Neutron deleted interface 185ea4de-3b36-467d-a112-f439b722671d; detaching it from the instance and deleting it from the info cache [ 757.942453] env[62385]: DEBUG nova.network.neutron [req-d8e1a38e-7f01-452e-bb0a-aefb8fceb303 req-3dd52a16-4910-4544-9c4e-0bd01ab792dd service nova] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.027778] env[62385]: DEBUG nova.compute.manager [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 758.039101] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 0ba32627-c646-477d-9062-b84624d01e48] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 758.058616] env[62385]: DEBUG nova.virt.hardware [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 758.059116] env[62385]: DEBUG nova.virt.hardware [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 758.059381] env[62385]: DEBUG nova.virt.hardware [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 758.059724] env[62385]: DEBUG nova.virt.hardware [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 758.060008] env[62385]: DEBUG nova.virt.hardware [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 758.060368] env[62385]: DEBUG nova.virt.hardware [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 758.060643] env[62385]: DEBUG nova.virt.hardware [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 758.061409] env[62385]: DEBUG nova.virt.hardware [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 758.061507] env[62385]: DEBUG nova.virt.hardware [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 758.061867] env[62385]: DEBUG nova.virt.hardware [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 758.062475] env[62385]: DEBUG nova.virt.hardware [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 758.064268] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79b35d4b-b0f9-4d6c-8227-a6f3cdd8cfd2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.077997] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2696c891-3a44-426a-900c-77c8cba2653f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.102114] env[62385]: DEBUG nova.scheduler.client.report [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 758.420617] env[62385]: DEBUG nova.network.neutron [-] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.445075] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cb07837b-bb22-41a6-a24e-ef5a1aab76e5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.455079] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8103065-5a11-4cb3-aaf3-450f9201ee9b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.483801] env[62385]: DEBUG nova.compute.manager [req-d8e1a38e-7f01-452e-bb0a-aefb8fceb303 req-3dd52a16-4910-4544-9c4e-0bd01ab792dd service nova] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Detach interface failed, port_id=185ea4de-3b36-467d-a112-f439b722671d, reason: Instance 42ecf594-dff7-4af3-ac56-5bbc45b7a192 could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 758.542403] env[62385]: DEBUG nova.compute.manager [req-2bbd705a-a7b2-4fe8-b415-0a9b1d7cb658 req-2de52dbf-b302-4a0e-b21f-3987d673efb2 service nova] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Received event network-vif-plugged-81b5c1c2-0651-4034-87db-8faf51c3c7d5 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 758.542403] env[62385]: DEBUG oslo_concurrency.lockutils [req-2bbd705a-a7b2-4fe8-b415-0a9b1d7cb658 req-2de52dbf-b302-4a0e-b21f-3987d673efb2 service nova] Acquiring lock "a5c4afc6-38a6-4815-8ec4-cc01c24489bf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.542834] env[62385]: DEBUG oslo_concurrency.lockutils [req-2bbd705a-a7b2-4fe8-b415-0a9b1d7cb658 req-2de52dbf-b302-4a0e-b21f-3987d673efb2 service nova] Lock "a5c4afc6-38a6-4815-8ec4-cc01c24489bf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 758.542834] env[62385]: DEBUG oslo_concurrency.lockutils [req-2bbd705a-a7b2-4fe8-b415-0a9b1d7cb658 req-2de52dbf-b302-4a0e-b21f-3987d673efb2 service nova] Lock "a5c4afc6-38a6-4815-8ec4-cc01c24489bf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 758.542834] env[62385]: DEBUG nova.compute.manager [req-2bbd705a-a7b2-4fe8-b415-0a9b1d7cb658 req-2de52dbf-b302-4a0e-b21f-3987d673efb2 service nova] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] No waiting events found dispatching network-vif-plugged-81b5c1c2-0651-4034-87db-8faf51c3c7d5 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 758.542997] env[62385]: WARNING nova.compute.manager [req-2bbd705a-a7b2-4fe8-b415-0a9b1d7cb658 req-2de52dbf-b302-4a0e-b21f-3987d673efb2 service nova] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Received unexpected event network-vif-plugged-81b5c1c2-0651-4034-87db-8faf51c3c7d5 for instance with vm_state building and task_state spawning. [ 758.543706] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: e514a72d-6b4d-46f9-935b-3535bed4afc2] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 758.607576] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.362s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 758.607760] env[62385]: DEBUG nova.compute.manager [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 758.610601] env[62385]: DEBUG oslo_concurrency.lockutils [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.917s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 758.611864] env[62385]: INFO nova.compute.claims [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 758.655829] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0e43bfd4-0da1-49c6-a0cb-aa2b4014286e tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Acquiring lock "92057af7-28a3-4643-9cda-d3d868d0cfc9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.656113] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0e43bfd4-0da1-49c6-a0cb-aa2b4014286e tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Lock "92057af7-28a3-4643-9cda-d3d868d0cfc9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 758.657031] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0e43bfd4-0da1-49c6-a0cb-aa2b4014286e tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Acquiring lock "92057af7-28a3-4643-9cda-d3d868d0cfc9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.657031] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0e43bfd4-0da1-49c6-a0cb-aa2b4014286e tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Lock "92057af7-28a3-4643-9cda-d3d868d0cfc9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 758.657031] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0e43bfd4-0da1-49c6-a0cb-aa2b4014286e tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Lock "92057af7-28a3-4643-9cda-d3d868d0cfc9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 758.659386] env[62385]: INFO nova.compute.manager [None req-0e43bfd4-0da1-49c6-a0cb-aa2b4014286e tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Terminating instance [ 758.662330] env[62385]: DEBUG nova.compute.manager [None req-0e43bfd4-0da1-49c6-a0cb-aa2b4014286e tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 758.662528] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0e43bfd4-0da1-49c6-a0cb-aa2b4014286e tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 758.663372] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec8bf328-b632-47b0-a466-7b66c0bfcf5f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.672580] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e43bfd4-0da1-49c6-a0cb-aa2b4014286e tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 758.672855] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-19514cab-1b84-4000-9580-abdfb19c1c26 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.680936] env[62385]: DEBUG oslo_vmware.api [None req-0e43bfd4-0da1-49c6-a0cb-aa2b4014286e tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for the task: (returnval){ [ 758.680936] env[62385]: value = "task-1205603" [ 758.680936] env[62385]: _type = "Task" [ 758.680936] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.690053] env[62385]: DEBUG oslo_vmware.api [None req-0e43bfd4-0da1-49c6-a0cb-aa2b4014286e tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205603, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.730994] env[62385]: DEBUG nova.network.neutron [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Successfully updated port: 81b5c1c2-0651-4034-87db-8faf51c3c7d5 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 758.922992] env[62385]: INFO nova.compute.manager [-] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Took 1.24 seconds to deallocate network for instance. [ 759.046718] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 759.046988] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Cleaning up deleted instances with incomplete migration {{(pid=62385) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 759.118540] env[62385]: DEBUG nova.compute.utils [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 759.119908] env[62385]: DEBUG nova.compute.manager [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 759.120194] env[62385]: DEBUG nova.network.neutron [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 759.159439] env[62385]: DEBUG nova.policy [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dfccdfd301334545aa0eaa796a955c43', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2ebc1c1ed0bb404b89e59325535edecd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 759.191345] env[62385]: DEBUG oslo_vmware.api [None req-0e43bfd4-0da1-49c6-a0cb-aa2b4014286e tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205603, 'name': PowerOffVM_Task, 'duration_secs': 0.188307} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.191628] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e43bfd4-0da1-49c6-a0cb-aa2b4014286e tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 759.191795] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0e43bfd4-0da1-49c6-a0cb-aa2b4014286e tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 759.192053] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e5d4f82f-751e-4047-99ff-6c91fef4a922 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.234398] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Acquiring lock "refresh_cache-a5c4afc6-38a6-4815-8ec4-cc01c24489bf" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.234398] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Acquired lock "refresh_cache-a5c4afc6-38a6-4815-8ec4-cc01c24489bf" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.234398] env[62385]: DEBUG nova.network.neutron [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 759.257027] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0e43bfd4-0da1-49c6-a0cb-aa2b4014286e tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 759.257027] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0e43bfd4-0da1-49c6-a0cb-aa2b4014286e tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Deleting contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 759.257027] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e43bfd4-0da1-49c6-a0cb-aa2b4014286e tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Deleting the datastore file [datastore2] 92057af7-28a3-4643-9cda-d3d868d0cfc9 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 759.257027] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-88d5b341-6abd-4fae-8c4f-3859af98766b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.265392] env[62385]: DEBUG oslo_vmware.api [None req-0e43bfd4-0da1-49c6-a0cb-aa2b4014286e tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for the task: (returnval){ [ 759.265392] env[62385]: value = "task-1205605" [ 759.265392] env[62385]: _type = "Task" [ 759.265392] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.273447] env[62385]: DEBUG oslo_vmware.api [None req-0e43bfd4-0da1-49c6-a0cb-aa2b4014286e tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205605, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.430990] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c3bb0442-402a-44c3-a5ec-63b1cb851d77 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 759.541098] env[62385]: DEBUG nova.network.neutron [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Successfully created port: 30e19299-99d0-4c98-8615-78f2f5f76dd5 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 759.552397] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 759.627603] env[62385]: DEBUG nova.compute.manager [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 759.783020] env[62385]: DEBUG oslo_vmware.api [None req-0e43bfd4-0da1-49c6-a0cb-aa2b4014286e tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205605, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.143478} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.783020] env[62385]: DEBUG nova.network.neutron [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 759.783020] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e43bfd4-0da1-49c6-a0cb-aa2b4014286e tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 759.783020] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0e43bfd4-0da1-49c6-a0cb-aa2b4014286e tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Deleted contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 759.783020] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0e43bfd4-0da1-49c6-a0cb-aa2b4014286e tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 759.783550] env[62385]: INFO nova.compute.manager [None req-0e43bfd4-0da1-49c6-a0cb-aa2b4014286e tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Took 1.12 seconds to destroy the instance on the hypervisor. [ 759.783550] env[62385]: DEBUG oslo.service.loopingcall [None req-0e43bfd4-0da1-49c6-a0cb-aa2b4014286e tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 759.783550] env[62385]: DEBUG nova.compute.manager [-] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 759.783550] env[62385]: DEBUG nova.network.neutron [-] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 759.926582] env[62385]: DEBUG nova.network.neutron [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Updating instance_info_cache with network_info: [{"id": "81b5c1c2-0651-4034-87db-8faf51c3c7d5", "address": "fa:16:3e:95:04:51", "network": {"id": "ddc6e4c3-c34c-4777-964b-3680c1226a5b", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-1290413408-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f08231c6ff14924a8e51b77ad84bdb9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81b5c1c2-06", "ovs_interfaceid": "81b5c1c2-0651-4034-87db-8faf51c3c7d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.036876] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3cdd0e4-dce0-45cc-aa5b-a22bc3e2317b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.045613] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c61fc897-e7f8-4dcf-833b-be84bc2c156e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.089945] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-705db17b-fc0f-485d-89ab-c2e0490e533b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.099211] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5058a5fe-2bee-4c36-b462-1b795791f96a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.116951] env[62385]: DEBUG nova.compute.provider_tree [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 760.428709] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Releasing lock "refresh_cache-a5c4afc6-38a6-4815-8ec4-cc01c24489bf" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.429101] env[62385]: DEBUG nova.compute.manager [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Instance network_info: |[{"id": "81b5c1c2-0651-4034-87db-8faf51c3c7d5", "address": "fa:16:3e:95:04:51", "network": {"id": "ddc6e4c3-c34c-4777-964b-3680c1226a5b", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-1290413408-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f08231c6ff14924a8e51b77ad84bdb9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81b5c1c2-06", "ovs_interfaceid": "81b5c1c2-0651-4034-87db-8faf51c3c7d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 760.429763] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:95:04:51', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '225b6979-9329-403b-91fa-138bd41f6e83', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '81b5c1c2-0651-4034-87db-8faf51c3c7d5', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 760.437753] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Creating folder: Project (6f08231c6ff14924a8e51b77ad84bdb9). Parent ref: group-v261107. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 760.438348] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5a649da2-5d2d-4ccb-892b-3be835ce626e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.452063] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Created folder: Project (6f08231c6ff14924a8e51b77ad84bdb9) in parent group-v261107. [ 760.452290] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Creating folder: Instances. Parent ref: group-v261151. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 760.452578] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-34ef26b3-7e8b-4080-912f-dafe16466b79 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.464049] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Created folder: Instances in parent group-v261151. [ 760.464334] env[62385]: DEBUG oslo.service.loopingcall [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 760.464547] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 760.464759] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fb93c917-76c6-403d-b085-f8216601f8bd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.487711] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 760.487711] env[62385]: value = "task-1205608" [ 760.487711] env[62385]: _type = "Task" [ 760.487711] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.496762] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205608, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.611309] env[62385]: DEBUG nova.compute.manager [req-364d42e1-30f1-488d-b85a-4a5322e3038a req-b5cc6ec4-38df-49f8-aac1-b97090dbf587 service nova] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Received event network-changed-81b5c1c2-0651-4034-87db-8faf51c3c7d5 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 760.611550] env[62385]: DEBUG nova.compute.manager [req-364d42e1-30f1-488d-b85a-4a5322e3038a req-b5cc6ec4-38df-49f8-aac1-b97090dbf587 service nova] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Refreshing instance network info cache due to event network-changed-81b5c1c2-0651-4034-87db-8faf51c3c7d5. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 760.611685] env[62385]: DEBUG oslo_concurrency.lockutils [req-364d42e1-30f1-488d-b85a-4a5322e3038a req-b5cc6ec4-38df-49f8-aac1-b97090dbf587 service nova] Acquiring lock "refresh_cache-a5c4afc6-38a6-4815-8ec4-cc01c24489bf" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.611827] env[62385]: DEBUG oslo_concurrency.lockutils [req-364d42e1-30f1-488d-b85a-4a5322e3038a req-b5cc6ec4-38df-49f8-aac1-b97090dbf587 service nova] Acquired lock "refresh_cache-a5c4afc6-38a6-4815-8ec4-cc01c24489bf" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.611987] env[62385]: DEBUG nova.network.neutron [req-364d42e1-30f1-488d-b85a-4a5322e3038a req-b5cc6ec4-38df-49f8-aac1-b97090dbf587 service nova] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Refreshing network info cache for port 81b5c1c2-0651-4034-87db-8faf51c3c7d5 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 760.619465] env[62385]: DEBUG nova.scheduler.client.report [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 760.636629] env[62385]: DEBUG nova.compute.manager [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 760.663043] env[62385]: DEBUG nova.virt.hardware [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 760.663301] env[62385]: DEBUG nova.virt.hardware [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 760.663462] env[62385]: DEBUG nova.virt.hardware [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 760.663643] env[62385]: DEBUG nova.virt.hardware [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 760.663791] env[62385]: DEBUG nova.virt.hardware [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 760.663939] env[62385]: DEBUG nova.virt.hardware [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 760.664252] env[62385]: DEBUG nova.virt.hardware [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 760.664426] env[62385]: DEBUG nova.virt.hardware [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 760.664596] env[62385]: DEBUG nova.virt.hardware [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 760.664760] env[62385]: DEBUG nova.virt.hardware [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 760.664928] env[62385]: DEBUG nova.virt.hardware [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 760.665853] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25635ff1-36d8-4a8a-ad5b-c0af8022b0a8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.676443] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f037e00b-957f-43cf-9a0d-87f2e6631c02 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.914052] env[62385]: DEBUG nova.network.neutron [-] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.998052] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205608, 'name': CreateVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.124605] env[62385]: DEBUG oslo_concurrency.lockutils [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.514s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.125121] env[62385]: DEBUG nova.compute.manager [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 761.127995] env[62385]: DEBUG oslo_concurrency.lockutils [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.495s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.129774] env[62385]: INFO nova.compute.claims [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 761.417168] env[62385]: INFO nova.compute.manager [-] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Took 1.63 seconds to deallocate network for instance. [ 761.500072] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205608, 'name': CreateVM_Task, 'duration_secs': 0.514874} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.500072] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 761.500261] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.501029] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.501029] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 761.501406] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd5cba91-f927-420a-9ff4-6c0e9fdf02e0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.506466] env[62385]: DEBUG oslo_vmware.api [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Waiting for the task: (returnval){ [ 761.506466] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521b5337-61a8-77d8-6942-d22f005fab85" [ 761.506466] env[62385]: _type = "Task" [ 761.506466] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.515570] env[62385]: DEBUG oslo_vmware.api [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521b5337-61a8-77d8-6942-d22f005fab85, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.540527] env[62385]: DEBUG nova.network.neutron [req-364d42e1-30f1-488d-b85a-4a5322e3038a req-b5cc6ec4-38df-49f8-aac1-b97090dbf587 service nova] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Updated VIF entry in instance network info cache for port 81b5c1c2-0651-4034-87db-8faf51c3c7d5. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 761.540898] env[62385]: DEBUG nova.network.neutron [req-364d42e1-30f1-488d-b85a-4a5322e3038a req-b5cc6ec4-38df-49f8-aac1-b97090dbf587 service nova] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Updating instance_info_cache with network_info: [{"id": "81b5c1c2-0651-4034-87db-8faf51c3c7d5", "address": "fa:16:3e:95:04:51", "network": {"id": "ddc6e4c3-c34c-4777-964b-3680c1226a5b", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-1290413408-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f08231c6ff14924a8e51b77ad84bdb9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81b5c1c2-06", "ovs_interfaceid": "81b5c1c2-0651-4034-87db-8faf51c3c7d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.634436] env[62385]: DEBUG nova.compute.utils [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 761.637962] env[62385]: DEBUG nova.compute.manager [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 761.638160] env[62385]: DEBUG nova.network.neutron [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 761.689770] env[62385]: DEBUG nova.policy [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '12b0c7b5283e4128a9e85d3edd52f2ba', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '395adff2d8204123b16bf9e8f5c04d05', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 761.820947] env[62385]: DEBUG nova.compute.manager [req-acc3c795-6411-47d9-90a7-a71ec4153dcf req-9942d6e1-7202-4596-a986-2068e76d965e service nova] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Received event network-vif-plugged-30e19299-99d0-4c98-8615-78f2f5f76dd5 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 761.821215] env[62385]: DEBUG oslo_concurrency.lockutils [req-acc3c795-6411-47d9-90a7-a71ec4153dcf req-9942d6e1-7202-4596-a986-2068e76d965e service nova] Acquiring lock "e73ecefc-fed3-4ccd-88d3-399a4b72bb4b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.821549] env[62385]: DEBUG oslo_concurrency.lockutils [req-acc3c795-6411-47d9-90a7-a71ec4153dcf req-9942d6e1-7202-4596-a986-2068e76d965e service nova] Lock "e73ecefc-fed3-4ccd-88d3-399a4b72bb4b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.821725] env[62385]: DEBUG oslo_concurrency.lockutils [req-acc3c795-6411-47d9-90a7-a71ec4153dcf req-9942d6e1-7202-4596-a986-2068e76d965e service nova] Lock "e73ecefc-fed3-4ccd-88d3-399a4b72bb4b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.821895] env[62385]: DEBUG nova.compute.manager [req-acc3c795-6411-47d9-90a7-a71ec4153dcf req-9942d6e1-7202-4596-a986-2068e76d965e service nova] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] No waiting events found dispatching network-vif-plugged-30e19299-99d0-4c98-8615-78f2f5f76dd5 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 761.822071] env[62385]: WARNING nova.compute.manager [req-acc3c795-6411-47d9-90a7-a71ec4153dcf req-9942d6e1-7202-4596-a986-2068e76d965e service nova] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Received unexpected event network-vif-plugged-30e19299-99d0-4c98-8615-78f2f5f76dd5 for instance with vm_state building and task_state spawning. [ 761.924757] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0e43bfd4-0da1-49c6-a0cb-aa2b4014286e tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.016849] env[62385]: DEBUG oslo_vmware.api [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521b5337-61a8-77d8-6942-d22f005fab85, 'name': SearchDatastore_Task, 'duration_secs': 0.010861} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.017323] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.018204] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 762.018204] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.018204] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.018204] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 762.018670] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c4f19c27-3b5e-44aa-a686-8c2bbfbd4a1e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.028368] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 762.028368] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 762.028368] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a689687e-a2b6-446d-88b6-01f0d27c90b9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.033491] env[62385]: DEBUG oslo_vmware.api [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Waiting for the task: (returnval){ [ 762.033491] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521c2a95-092d-e9a3-ffb7-ceb89705e9a5" [ 762.033491] env[62385]: _type = "Task" [ 762.033491] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.042073] env[62385]: DEBUG oslo_vmware.api [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521c2a95-092d-e9a3-ffb7-ceb89705e9a5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.042867] env[62385]: DEBUG nova.network.neutron [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Successfully created port: b838271e-07e8-4880-980a-703c1bbd4da5 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 762.048210] env[62385]: DEBUG oslo_concurrency.lockutils [req-364d42e1-30f1-488d-b85a-4a5322e3038a req-b5cc6ec4-38df-49f8-aac1-b97090dbf587 service nova] Releasing lock "refresh_cache-a5c4afc6-38a6-4815-8ec4-cc01c24489bf" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.048468] env[62385]: DEBUG nova.compute.manager [req-364d42e1-30f1-488d-b85a-4a5322e3038a req-b5cc6ec4-38df-49f8-aac1-b97090dbf587 service nova] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Received event network-vif-deleted-de1e1c07-3ca9-4e4f-ae07-8e43ed762fc4 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 762.048669] env[62385]: INFO nova.compute.manager [req-364d42e1-30f1-488d-b85a-4a5322e3038a req-b5cc6ec4-38df-49f8-aac1-b97090dbf587 service nova] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Neutron deleted interface de1e1c07-3ca9-4e4f-ae07-8e43ed762fc4; detaching it from the instance and deleting it from the info cache [ 762.048843] env[62385]: DEBUG nova.network.neutron [req-364d42e1-30f1-488d-b85a-4a5322e3038a req-b5cc6ec4-38df-49f8-aac1-b97090dbf587 service nova] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.089755] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 762.090098] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 762.090323] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Starting heal instance info cache {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 762.090483] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Rebuilding the list of instances to heal {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 762.139221] env[62385]: DEBUG nova.compute.manager [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 762.167249] env[62385]: DEBUG nova.network.neutron [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Successfully updated port: 30e19299-99d0-4c98-8615-78f2f5f76dd5 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 762.500558] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a120eb4a-d195-4c37-b1f7-521b58b8c6ca {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.508427] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a27b1df3-3df3-48ae-8998-c30bd09531ef {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.541451] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6f45bba-3d8c-43d3-932d-0888202b084a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.553026] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bdd6e81-33b1-4331-80bf-8317c3165f6d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.556530] env[62385]: DEBUG oslo_vmware.api [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521c2a95-092d-e9a3-ffb7-ceb89705e9a5, 'name': SearchDatastore_Task, 'duration_secs': 0.010365} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.556742] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-60318764-7aa4-409d-bf2c-5ef662eb3da2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.559419] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d9d079a3-e2be-452e-9d61-c3784b1b17dc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.569628] env[62385]: DEBUG nova.compute.provider_tree [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 762.576810] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2b3216d-cc61-414c-8d2c-0959de69f97a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.587012] env[62385]: DEBUG oslo_vmware.api [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Waiting for the task: (returnval){ [ 762.587012] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]527646c2-9597-435e-162b-39971c55e84b" [ 762.587012] env[62385]: _type = "Task" [ 762.587012] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.596805] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Skipping network cache update for instance because it is Building. {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 762.596958] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Skipping network cache update for instance because it is Building. {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 762.597101] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Skipping network cache update for instance because it is Building. {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 762.597229] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Skipping network cache update for instance because it is Building. {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 762.613212] env[62385]: DEBUG nova.compute.manager [req-364d42e1-30f1-488d-b85a-4a5322e3038a req-b5cc6ec4-38df-49f8-aac1-b97090dbf587 service nova] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Detach interface failed, port_id=de1e1c07-3ca9-4e4f-ae07-8e43ed762fc4, reason: Instance 92057af7-28a3-4643-9cda-d3d868d0cfc9 could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 762.613641] env[62385]: DEBUG oslo_vmware.api [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]527646c2-9597-435e-162b-39971c55e84b, 'name': SearchDatastore_Task, 'duration_secs': 0.011149} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.614204] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.614450] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] a5c4afc6-38a6-4815-8ec4-cc01c24489bf/a5c4afc6-38a6-4815-8ec4-cc01c24489bf.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 762.614699] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-299957a6-6939-4c4f-bac4-a3f925ac5a02 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.622623] env[62385]: DEBUG oslo_vmware.api [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Waiting for the task: (returnval){ [ 762.622623] env[62385]: value = "task-1205609" [ 762.622623] env[62385]: _type = "Task" [ 762.622623] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.631414] env[62385]: DEBUG oslo_vmware.api [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205609, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.672961] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Acquiring lock "refresh_cache-e73ecefc-fed3-4ccd-88d3-399a4b72bb4b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.673230] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Acquired lock "refresh_cache-e73ecefc-fed3-4ccd-88d3-399a4b72bb4b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.673281] env[62385]: DEBUG nova.network.neutron [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 762.691872] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquiring lock "refresh_cache-1ddb4d6d-3872-49db-bb40-e21721241e89" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.693061] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquired lock "refresh_cache-1ddb4d6d-3872-49db-bb40-e21721241e89" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.693061] env[62385]: DEBUG nova.network.neutron [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Forcefully refreshing network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 762.693061] env[62385]: DEBUG nova.objects.instance [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lazy-loading 'info_cache' on Instance uuid 1ddb4d6d-3872-49db-bb40-e21721241e89 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 762.942321] env[62385]: DEBUG nova.compute.manager [req-65a3b950-3517-4ab3-a6ff-f1d7484b828b req-ec8bb8f3-72b7-46f9-91ab-de4e3c492e5b service nova] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Received event network-changed-30e19299-99d0-4c98-8615-78f2f5f76dd5 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 762.942519] env[62385]: DEBUG nova.compute.manager [req-65a3b950-3517-4ab3-a6ff-f1d7484b828b req-ec8bb8f3-72b7-46f9-91ab-de4e3c492e5b service nova] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Refreshing instance network info cache due to event network-changed-30e19299-99d0-4c98-8615-78f2f5f76dd5. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 762.942709] env[62385]: DEBUG oslo_concurrency.lockutils [req-65a3b950-3517-4ab3-a6ff-f1d7484b828b req-ec8bb8f3-72b7-46f9-91ab-de4e3c492e5b service nova] Acquiring lock "refresh_cache-e73ecefc-fed3-4ccd-88d3-399a4b72bb4b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.075020] env[62385]: DEBUG nova.scheduler.client.report [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 763.132905] env[62385]: DEBUG oslo_vmware.api [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205609, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.446479} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.132905] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] a5c4afc6-38a6-4815-8ec4-cc01c24489bf/a5c4afc6-38a6-4815-8ec4-cc01c24489bf.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 763.133248] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 763.133475] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4757c8f8-75ac-482c-afc6-1b6a1b2ad816 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.140484] env[62385]: DEBUG oslo_vmware.api [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Waiting for the task: (returnval){ [ 763.140484] env[62385]: value = "task-1205610" [ 763.140484] env[62385]: _type = "Task" [ 763.140484] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.149203] env[62385]: DEBUG nova.compute.manager [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 763.151126] env[62385]: DEBUG oslo_vmware.api [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205610, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.175699] env[62385]: DEBUG nova.virt.hardware [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 763.176037] env[62385]: DEBUG nova.virt.hardware [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 763.176221] env[62385]: DEBUG nova.virt.hardware [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 763.176471] env[62385]: DEBUG nova.virt.hardware [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 763.176632] env[62385]: DEBUG nova.virt.hardware [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 763.176780] env[62385]: DEBUG nova.virt.hardware [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 763.176991] env[62385]: DEBUG nova.virt.hardware [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 763.177181] env[62385]: DEBUG nova.virt.hardware [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 763.177461] env[62385]: DEBUG nova.virt.hardware [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 763.177638] env[62385]: DEBUG nova.virt.hardware [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 763.177903] env[62385]: DEBUG nova.virt.hardware [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 763.180576] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5487b765-f32f-485a-8544-c1aedb2706fb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.189329] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4681086-e1fc-4ce8-b040-4d9078eeda5d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.229919] env[62385]: DEBUG nova.network.neutron [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 763.417844] env[62385]: DEBUG nova.network.neutron [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Updating instance_info_cache with network_info: [{"id": "30e19299-99d0-4c98-8615-78f2f5f76dd5", "address": "fa:16:3e:a9:58:3d", "network": {"id": "63367e0f-9855-478b-a546-5f0cf8190519", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-21852824-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ebc1c1ed0bb404b89e59325535edecd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "778b9a40-d603-4765-ac88-bd6d42c457a2", "external-id": "nsx-vlan-transportzone-114", "segmentation_id": 114, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30e19299-99", "ovs_interfaceid": "30e19299-99d0-4c98-8615-78f2f5f76dd5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.577587] env[62385]: DEBUG oslo_concurrency.lockutils [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.450s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.578229] env[62385]: DEBUG nova.compute.manager [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 763.580778] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.863s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 763.582150] env[62385]: INFO nova.compute.claims [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 763.652805] env[62385]: DEBUG oslo_vmware.api [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205610, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068867} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.653299] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 763.654166] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87f45ba5-787f-489a-95de-1cc2905e97ac {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.676381] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Reconfiguring VM instance instance-00000035 to attach disk [datastore1] a5c4afc6-38a6-4815-8ec4-cc01c24489bf/a5c4afc6-38a6-4815-8ec4-cc01c24489bf.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 763.676685] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-436228e5-1830-43ad-81e2-bdacd08beabc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.703830] env[62385]: DEBUG oslo_vmware.api [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Waiting for the task: (returnval){ [ 763.703830] env[62385]: value = "task-1205611" [ 763.703830] env[62385]: _type = "Task" [ 763.703830] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.708201] env[62385]: DEBUG nova.network.neutron [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Successfully updated port: b838271e-07e8-4880-980a-703c1bbd4da5 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 763.718159] env[62385]: DEBUG oslo_vmware.api [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205611, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.921045] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Releasing lock "refresh_cache-e73ecefc-fed3-4ccd-88d3-399a4b72bb4b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.921045] env[62385]: DEBUG nova.compute.manager [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Instance network_info: |[{"id": "30e19299-99d0-4c98-8615-78f2f5f76dd5", "address": "fa:16:3e:a9:58:3d", "network": {"id": "63367e0f-9855-478b-a546-5f0cf8190519", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-21852824-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ebc1c1ed0bb404b89e59325535edecd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "778b9a40-d603-4765-ac88-bd6d42c457a2", "external-id": "nsx-vlan-transportzone-114", "segmentation_id": 114, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30e19299-99", "ovs_interfaceid": "30e19299-99d0-4c98-8615-78f2f5f76dd5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 763.921266] env[62385]: DEBUG oslo_concurrency.lockutils [req-65a3b950-3517-4ab3-a6ff-f1d7484b828b req-ec8bb8f3-72b7-46f9-91ab-de4e3c492e5b service nova] Acquired lock "refresh_cache-e73ecefc-fed3-4ccd-88d3-399a4b72bb4b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.921266] env[62385]: DEBUG nova.network.neutron [req-65a3b950-3517-4ab3-a6ff-f1d7484b828b req-ec8bb8f3-72b7-46f9-91ab-de4e3c492e5b service nova] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Refreshing network info cache for port 30e19299-99d0-4c98-8615-78f2f5f76dd5 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 763.922375] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a9:58:3d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '778b9a40-d603-4765-ac88-bd6d42c457a2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '30e19299-99d0-4c98-8615-78f2f5f76dd5', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 763.930959] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Creating folder: Project (2ebc1c1ed0bb404b89e59325535edecd). Parent ref: group-v261107. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 763.934177] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-21097a2d-75f9-4396-bf37-3e059242af70 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.948282] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Created folder: Project (2ebc1c1ed0bb404b89e59325535edecd) in parent group-v261107. [ 763.948457] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Creating folder: Instances. Parent ref: group-v261154. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 763.948768] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e8b3d0dc-9118-4723-8550-c6fbdf35a988 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.959259] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Created folder: Instances in parent group-v261154. [ 763.959525] env[62385]: DEBUG oslo.service.loopingcall [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 763.959722] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 763.959932] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6cb072d8-2561-4f09-9ca8-771d806e93b4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.983111] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 763.983111] env[62385]: value = "task-1205614" [ 763.983111] env[62385]: _type = "Task" [ 763.983111] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.991793] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205614, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.088208] env[62385]: DEBUG nova.compute.utils [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 764.091653] env[62385]: DEBUG nova.compute.manager [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 764.091879] env[62385]: DEBUG nova.network.neutron [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 764.199272] env[62385]: DEBUG nova.policy [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e1200cdf75b04233b4c3f4626d4e27cf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aea48968d4464ac0ab01c998c059107d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 764.213190] env[62385]: DEBUG oslo_vmware.api [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205611, 'name': ReconfigVM_Task, 'duration_secs': 0.311038} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.213190] env[62385]: DEBUG oslo_concurrency.lockutils [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Acquiring lock "refresh_cache-65bb1d14-84b2-4e75-acdc-dc674a035101" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.213190] env[62385]: DEBUG oslo_concurrency.lockutils [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Acquired lock "refresh_cache-65bb1d14-84b2-4e75-acdc-dc674a035101" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.213190] env[62385]: DEBUG nova.network.neutron [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 764.214319] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Reconfigured VM instance instance-00000035 to attach disk [datastore1] a5c4afc6-38a6-4815-8ec4-cc01c24489bf/a5c4afc6-38a6-4815-8ec4-cc01c24489bf.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 764.214319] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-869f79b3-3488-4d04-aa1a-6ac47db354cd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.227018] env[62385]: DEBUG oslo_vmware.api [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Waiting for the task: (returnval){ [ 764.227018] env[62385]: value = "task-1205615" [ 764.227018] env[62385]: _type = "Task" [ 764.227018] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.239318] env[62385]: DEBUG oslo_vmware.api [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205615, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.347915] env[62385]: DEBUG nova.network.neutron [req-65a3b950-3517-4ab3-a6ff-f1d7484b828b req-ec8bb8f3-72b7-46f9-91ab-de4e3c492e5b service nova] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Updated VIF entry in instance network info cache for port 30e19299-99d0-4c98-8615-78f2f5f76dd5. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 764.348376] env[62385]: DEBUG nova.network.neutron [req-65a3b950-3517-4ab3-a6ff-f1d7484b828b req-ec8bb8f3-72b7-46f9-91ab-de4e3c492e5b service nova] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Updating instance_info_cache with network_info: [{"id": "30e19299-99d0-4c98-8615-78f2f5f76dd5", "address": "fa:16:3e:a9:58:3d", "network": {"id": "63367e0f-9855-478b-a546-5f0cf8190519", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-21852824-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ebc1c1ed0bb404b89e59325535edecd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "778b9a40-d603-4765-ac88-bd6d42c457a2", "external-id": "nsx-vlan-transportzone-114", "segmentation_id": 114, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30e19299-99", "ovs_interfaceid": "30e19299-99d0-4c98-8615-78f2f5f76dd5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.461097] env[62385]: DEBUG nova.network.neutron [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Updating instance_info_cache with network_info: [{"id": "7965d15d-61f5-4fbb-8e38-53e0207a6c58", "address": "fa:16:3e:fb:f0:ae", "network": {"id": "1d0e9098-63d3-4df5-a80a-06f9a5d90db6", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-303269893-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72549cb052314f149111e4e24ffb6656", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "572b7281-aad3-45fa-9cb2-fc1c70569948", "external-id": "nsx-vlan-transportzone-722", "segmentation_id": 722, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7965d15d-61", "ovs_interfaceid": "7965d15d-61f5-4fbb-8e38-53e0207a6c58", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.496044] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205614, 'name': CreateVM_Task, 'duration_secs': 0.363759} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.496044] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 764.500017] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.500017] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.500017] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 764.500017] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a181f0d-cb1e-48ba-8e86-23a18c7daf40 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.503326] env[62385]: DEBUG oslo_vmware.api [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Waiting for the task: (returnval){ [ 764.503326] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5200fe8c-1c13-f5bc-643b-9adb149f0515" [ 764.503326] env[62385]: _type = "Task" [ 764.503326] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.512328] env[62385]: DEBUG oslo_vmware.api [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5200fe8c-1c13-f5bc-643b-9adb149f0515, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.588192] env[62385]: DEBUG nova.network.neutron [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Successfully created port: bfb1f5cb-27f6-4422-8a40-f8f2d8a6520b {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 764.593022] env[62385]: DEBUG nova.compute.manager [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 764.741270] env[62385]: DEBUG oslo_vmware.api [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205615, 'name': Rename_Task, 'duration_secs': 0.158412} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.741590] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 764.741842] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7c804e2e-6f13-458f-a1f1-3c90bd268a50 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.750681] env[62385]: DEBUG oslo_vmware.api [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Waiting for the task: (returnval){ [ 764.750681] env[62385]: value = "task-1205616" [ 764.750681] env[62385]: _type = "Task" [ 764.750681] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.758069] env[62385]: DEBUG nova.network.neutron [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 764.764810] env[62385]: DEBUG oslo_vmware.api [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205616, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.851751] env[62385]: DEBUG oslo_concurrency.lockutils [req-65a3b950-3517-4ab3-a6ff-f1d7484b828b req-ec8bb8f3-72b7-46f9-91ab-de4e3c492e5b service nova] Releasing lock "refresh_cache-e73ecefc-fed3-4ccd-88d3-399a4b72bb4b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.895338] env[62385]: DEBUG nova.network.neutron [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Updating instance_info_cache with network_info: [{"id": "b838271e-07e8-4880-980a-703c1bbd4da5", "address": "fa:16:3e:f6:2c:f9", "network": {"id": "1c117704-eb0e-43fc-84fe-5e5059cbb310", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1332686438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "395adff2d8204123b16bf9e8f5c04d05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56834f67-27a8-43dc-bbc6-a74aaa08959b", "external-id": "nsx-vlan-transportzone-949", "segmentation_id": 949, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb838271e-07", "ovs_interfaceid": "b838271e-07e8-4880-980a-703c1bbd4da5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.963981] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Releasing lock "refresh_cache-1ddb4d6d-3872-49db-bb40-e21721241e89" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.964222] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Updated the network info_cache for instance {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 764.964421] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 764.964576] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 764.964721] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 764.964878] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 764.965049] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 764.965209] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 764.965334] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62385) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 764.965476] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager.update_available_resource {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 765.019919] env[62385]: DEBUG oslo_vmware.api [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5200fe8c-1c13-f5bc-643b-9adb149f0515, 'name': SearchDatastore_Task, 'duration_secs': 0.009696} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.020256] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.020508] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 765.020795] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.020984] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.021202] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 765.021472] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-40963e89-d650-44a0-b1c0-efdeaf168c42 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.030682] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 765.030877] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 765.031637] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27e6582c-e1eb-4f84-8ae2-c50e8ff332ca {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.037103] env[62385]: DEBUG oslo_vmware.api [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Waiting for the task: (returnval){ [ 765.037103] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52b5d500-c0e8-36c0-db39-8f369b552403" [ 765.037103] env[62385]: _type = "Task" [ 765.037103] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.042258] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16711148-b792-47ab-95d8-4b9aa5aed840 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.053287] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bce3d57-bdad-4f33-a9d2-72560c65cf25 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.056606] env[62385]: DEBUG oslo_vmware.api [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52b5d500-c0e8-36c0-db39-8f369b552403, 'name': SearchDatastore_Task, 'duration_secs': 0.00899} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.057679] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-23de9fd5-2fe4-48ae-9808-7c736ba30ce4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.091463] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d718cfe9-5f24-4a5a-b305-fab838e5500b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.094097] env[62385]: DEBUG oslo_vmware.api [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Waiting for the task: (returnval){ [ 765.094097] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c92402-1985-f926-05a3-b4bfd0d7978d" [ 765.094097] env[62385]: _type = "Task" [ 765.094097] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.103746] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06850c22-24d0-4cbc-b033-2e1cbae7788d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.111161] env[62385]: DEBUG oslo_vmware.api [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c92402-1985-f926-05a3-b4bfd0d7978d, 'name': SearchDatastore_Task, 'duration_secs': 0.010068} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.112225] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.112225] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] e73ecefc-fed3-4ccd-88d3-399a4b72bb4b/e73ecefc-fed3-4ccd-88d3-399a4b72bb4b.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 765.112467] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3be2f227-a764-42c1-9cba-f26ba268438d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.122529] env[62385]: DEBUG nova.compute.provider_tree [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 765.130035] env[62385]: DEBUG oslo_vmware.api [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Waiting for the task: (returnval){ [ 765.130035] env[62385]: value = "task-1205617" [ 765.130035] env[62385]: _type = "Task" [ 765.130035] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.142202] env[62385]: DEBUG oslo_vmware.api [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Task: {'id': task-1205617, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.249964] env[62385]: DEBUG nova.compute.manager [req-5c4b581c-53d4-4ae7-8d46-243aa9414b42 req-32303c31-2fc3-433a-83f0-6e9cbf4ccdbd service nova] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Received event network-vif-plugged-b838271e-07e8-4880-980a-703c1bbd4da5 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 765.251174] env[62385]: DEBUG oslo_concurrency.lockutils [req-5c4b581c-53d4-4ae7-8d46-243aa9414b42 req-32303c31-2fc3-433a-83f0-6e9cbf4ccdbd service nova] Acquiring lock "65bb1d14-84b2-4e75-acdc-dc674a035101-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.251174] env[62385]: DEBUG oslo_concurrency.lockutils [req-5c4b581c-53d4-4ae7-8d46-243aa9414b42 req-32303c31-2fc3-433a-83f0-6e9cbf4ccdbd service nova] Lock "65bb1d14-84b2-4e75-acdc-dc674a035101-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.251174] env[62385]: DEBUG oslo_concurrency.lockutils [req-5c4b581c-53d4-4ae7-8d46-243aa9414b42 req-32303c31-2fc3-433a-83f0-6e9cbf4ccdbd service nova] Lock "65bb1d14-84b2-4e75-acdc-dc674a035101-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.251174] env[62385]: DEBUG nova.compute.manager [req-5c4b581c-53d4-4ae7-8d46-243aa9414b42 req-32303c31-2fc3-433a-83f0-6e9cbf4ccdbd service nova] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] No waiting events found dispatching network-vif-plugged-b838271e-07e8-4880-980a-703c1bbd4da5 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 765.251578] env[62385]: WARNING nova.compute.manager [req-5c4b581c-53d4-4ae7-8d46-243aa9414b42 req-32303c31-2fc3-433a-83f0-6e9cbf4ccdbd service nova] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Received unexpected event network-vif-plugged-b838271e-07e8-4880-980a-703c1bbd4da5 for instance with vm_state building and task_state spawning. [ 765.251578] env[62385]: DEBUG nova.compute.manager [req-5c4b581c-53d4-4ae7-8d46-243aa9414b42 req-32303c31-2fc3-433a-83f0-6e9cbf4ccdbd service nova] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Received event network-changed-b838271e-07e8-4880-980a-703c1bbd4da5 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 765.251777] env[62385]: DEBUG nova.compute.manager [req-5c4b581c-53d4-4ae7-8d46-243aa9414b42 req-32303c31-2fc3-433a-83f0-6e9cbf4ccdbd service nova] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Refreshing instance network info cache due to event network-changed-b838271e-07e8-4880-980a-703c1bbd4da5. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 765.251874] env[62385]: DEBUG oslo_concurrency.lockutils [req-5c4b581c-53d4-4ae7-8d46-243aa9414b42 req-32303c31-2fc3-433a-83f0-6e9cbf4ccdbd service nova] Acquiring lock "refresh_cache-65bb1d14-84b2-4e75-acdc-dc674a035101" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.271151] env[62385]: DEBUG oslo_vmware.api [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205616, 'name': PowerOnVM_Task, 'duration_secs': 0.452163} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.271151] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 765.271151] env[62385]: INFO nova.compute.manager [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Took 7.24 seconds to spawn the instance on the hypervisor. [ 765.271345] env[62385]: DEBUG nova.compute.manager [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 765.272119] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10700a0a-60d9-45dc-acec-e6f07133cffe {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.401026] env[62385]: DEBUG oslo_concurrency.lockutils [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Releasing lock "refresh_cache-65bb1d14-84b2-4e75-acdc-dc674a035101" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.401026] env[62385]: DEBUG nova.compute.manager [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Instance network_info: |[{"id": "b838271e-07e8-4880-980a-703c1bbd4da5", "address": "fa:16:3e:f6:2c:f9", "network": {"id": "1c117704-eb0e-43fc-84fe-5e5059cbb310", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1332686438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "395adff2d8204123b16bf9e8f5c04d05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56834f67-27a8-43dc-bbc6-a74aaa08959b", "external-id": "nsx-vlan-transportzone-949", "segmentation_id": 949, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb838271e-07", "ovs_interfaceid": "b838271e-07e8-4880-980a-703c1bbd4da5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 765.401322] env[62385]: DEBUG oslo_concurrency.lockutils [req-5c4b581c-53d4-4ae7-8d46-243aa9414b42 req-32303c31-2fc3-433a-83f0-6e9cbf4ccdbd service nova] Acquired lock "refresh_cache-65bb1d14-84b2-4e75-acdc-dc674a035101" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.401322] env[62385]: DEBUG nova.network.neutron [req-5c4b581c-53d4-4ae7-8d46-243aa9414b42 req-32303c31-2fc3-433a-83f0-6e9cbf4ccdbd service nova] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Refreshing network info cache for port b838271e-07e8-4880-980a-703c1bbd4da5 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 765.401322] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f6:2c:f9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56834f67-27a8-43dc-bbc6-a74aaa08959b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b838271e-07e8-4880-980a-703c1bbd4da5', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 765.408970] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Creating folder: Project (395adff2d8204123b16bf9e8f5c04d05). Parent ref: group-v261107. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 765.409652] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9b0628bd-6c8e-417c-993f-8f60abd31e56 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.425088] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Created folder: Project (395adff2d8204123b16bf9e8f5c04d05) in parent group-v261107. [ 765.425355] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Creating folder: Instances. Parent ref: group-v261157. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 765.425617] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c01c93d8-f4ec-4a50-9110-7310b01bb3a2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.443398] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Created folder: Instances in parent group-v261157. [ 765.443715] env[62385]: DEBUG oslo.service.loopingcall [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 765.443927] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 765.444205] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9d72dfd6-87a7-4585-9676-0f2ce48efffe {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.469577] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.470066] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 765.470066] env[62385]: value = "task-1205620" [ 765.470066] env[62385]: _type = "Task" [ 765.470066] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.479922] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205620, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.609482] env[62385]: DEBUG nova.compute.manager [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 765.626122] env[62385]: DEBUG nova.scheduler.client.report [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 765.639371] env[62385]: DEBUG nova.virt.hardware [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 765.639613] env[62385]: DEBUG nova.virt.hardware [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 765.639771] env[62385]: DEBUG nova.virt.hardware [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 765.639951] env[62385]: DEBUG nova.virt.hardware [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 765.640110] env[62385]: DEBUG nova.virt.hardware [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 765.640282] env[62385]: DEBUG nova.virt.hardware [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 765.640493] env[62385]: DEBUG nova.virt.hardware [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 765.640653] env[62385]: DEBUG nova.virt.hardware [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 765.640818] env[62385]: DEBUG nova.virt.hardware [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 765.640979] env[62385]: DEBUG nova.virt.hardware [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 765.641170] env[62385]: DEBUG nova.virt.hardware [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 765.642359] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92adfd74-185b-41d1-b979-18a073534b39 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.647540] env[62385]: DEBUG oslo_vmware.api [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Task: {'id': task-1205617, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.653629] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21bb7e8d-1b82-45dd-a46d-6606552567f2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.793096] env[62385]: INFO nova.compute.manager [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Took 38.43 seconds to build instance. [ 765.982449] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205620, 'name': CreateVM_Task} progress is 25%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.093185] env[62385]: DEBUG nova.network.neutron [req-5c4b581c-53d4-4ae7-8d46-243aa9414b42 req-32303c31-2fc3-433a-83f0-6e9cbf4ccdbd service nova] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Updated VIF entry in instance network info cache for port b838271e-07e8-4880-980a-703c1bbd4da5. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 766.093185] env[62385]: DEBUG nova.network.neutron [req-5c4b581c-53d4-4ae7-8d46-243aa9414b42 req-32303c31-2fc3-433a-83f0-6e9cbf4ccdbd service nova] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Updating instance_info_cache with network_info: [{"id": "b838271e-07e8-4880-980a-703c1bbd4da5", "address": "fa:16:3e:f6:2c:f9", "network": {"id": "1c117704-eb0e-43fc-84fe-5e5059cbb310", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1332686438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "395adff2d8204123b16bf9e8f5c04d05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56834f67-27a8-43dc-bbc6-a74aaa08959b", "external-id": "nsx-vlan-transportzone-949", "segmentation_id": 949, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb838271e-07", "ovs_interfaceid": "b838271e-07e8-4880-980a-703c1bbd4da5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.134427] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.554s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.135034] env[62385]: DEBUG nova.compute.manager [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 766.142072] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.230s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 766.143637] env[62385]: INFO nova.compute.claims [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 766.152727] env[62385]: DEBUG oslo_vmware.api [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Task: {'id': task-1205617, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.514211} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.152824] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] e73ecefc-fed3-4ccd-88d3-399a4b72bb4b/e73ecefc-fed3-4ccd-88d3-399a4b72bb4b.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 766.153022] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 766.153282] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c9c901d5-98f2-420f-99bb-67c1a2343ab3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.160464] env[62385]: DEBUG oslo_vmware.api [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Waiting for the task: (returnval){ [ 766.160464] env[62385]: value = "task-1205621" [ 766.160464] env[62385]: _type = "Task" [ 766.160464] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.170107] env[62385]: DEBUG oslo_vmware.api [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Task: {'id': task-1205621, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.294811] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c44be613-d21f-476f-b660-59583067f043 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Lock "a5c4afc6-38a6-4815-8ec4-cc01c24489bf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 133.426s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.481136] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205620, 'name': CreateVM_Task, 'duration_secs': 0.704373} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.481325] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 766.481957] env[62385]: DEBUG oslo_concurrency.lockutils [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.482137] env[62385]: DEBUG oslo_concurrency.lockutils [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.482440] env[62385]: DEBUG oslo_concurrency.lockutils [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 766.482680] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-602f5d8d-d310-4ac6-ac88-b375533ff080 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.487526] env[62385]: DEBUG oslo_vmware.api [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Waiting for the task: (returnval){ [ 766.487526] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5286ec93-f480-5a33-c5e2-51331ceee401" [ 766.487526] env[62385]: _type = "Task" [ 766.487526] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.494997] env[62385]: DEBUG oslo_vmware.api [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5286ec93-f480-5a33-c5e2-51331ceee401, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.596785] env[62385]: DEBUG oslo_concurrency.lockutils [req-5c4b581c-53d4-4ae7-8d46-243aa9414b42 req-32303c31-2fc3-433a-83f0-6e9cbf4ccdbd service nova] Releasing lock "refresh_cache-65bb1d14-84b2-4e75-acdc-dc674a035101" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.648520] env[62385]: DEBUG nova.compute.utils [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 766.651846] env[62385]: DEBUG nova.compute.manager [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 766.652018] env[62385]: DEBUG nova.network.neutron [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 766.670765] env[62385]: DEBUG oslo_vmware.api [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Task: {'id': task-1205621, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.320217} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.671588] env[62385]: DEBUG nova.network.neutron [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Successfully updated port: bfb1f5cb-27f6-4422-8a40-f8f2d8a6520b {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 766.672702] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 766.674303] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d8713df-b946-4dd9-ae30-e3b92ddc2211 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.701023] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Reconfiguring VM instance instance-00000036 to attach disk [datastore1] e73ecefc-fed3-4ccd-88d3-399a4b72bb4b/e73ecefc-fed3-4ccd-88d3-399a4b72bb4b.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 766.702665] env[62385]: DEBUG nova.policy [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '50c2cb83af414c3f9f3ce647888ac454', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '06ca138339954db79c9bd6af2bcac5ff', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 766.704434] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3319fb03-0209-4129-b432-162b6f7ae5a8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.726463] env[62385]: DEBUG oslo_vmware.api [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Waiting for the task: (returnval){ [ 766.726463] env[62385]: value = "task-1205622" [ 766.726463] env[62385]: _type = "Task" [ 766.726463] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.737473] env[62385]: DEBUG oslo_vmware.api [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Task: {'id': task-1205622, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.797751] env[62385]: DEBUG nova.compute.manager [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 766.983183] env[62385]: DEBUG nova.network.neutron [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Successfully created port: 326dbe5f-be23-4610-bc79-079259f3a2c3 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 766.998686] env[62385]: DEBUG oslo_vmware.api [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5286ec93-f480-5a33-c5e2-51331ceee401, 'name': SearchDatastore_Task, 'duration_secs': 0.037342} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.999065] env[62385]: DEBUG oslo_concurrency.lockutils [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.000258] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 767.000697] env[62385]: DEBUG oslo_concurrency.lockutils [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.000775] env[62385]: DEBUG oslo_concurrency.lockutils [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.000915] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 767.001198] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bf46d406-6ca2-4c80-9667-76b3d2272acb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.012446] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 767.012622] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 767.013371] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d8897885-27b6-4a97-b442-db47cbe0349a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.019446] env[62385]: DEBUG oslo_vmware.api [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Waiting for the task: (returnval){ [ 767.019446] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]527573d3-ae2f-4bf4-0f26-849651f10488" [ 767.019446] env[62385]: _type = "Task" [ 767.019446] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.028406] env[62385]: DEBUG oslo_vmware.api [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]527573d3-ae2f-4bf4-0f26-849651f10488, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.152635] env[62385]: DEBUG nova.compute.manager [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 767.177020] env[62385]: DEBUG oslo_concurrency.lockutils [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "refresh_cache-f8a98710-9d02-4aa3-b6a5-34b7060ad62d" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.177020] env[62385]: DEBUG oslo_concurrency.lockutils [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquired lock "refresh_cache-f8a98710-9d02-4aa3-b6a5-34b7060ad62d" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.177020] env[62385]: DEBUG nova.network.neutron [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 767.238107] env[62385]: DEBUG oslo_vmware.api [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Task: {'id': task-1205622, 'name': ReconfigVM_Task, 'duration_secs': 0.288445} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.238375] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Reconfigured VM instance instance-00000036 to attach disk [datastore1] e73ecefc-fed3-4ccd-88d3-399a4b72bb4b/e73ecefc-fed3-4ccd-88d3-399a4b72bb4b.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 767.239200] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-36a9712b-68a6-41ca-b1a2-caf30819ec5a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.246605] env[62385]: DEBUG oslo_vmware.api [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Waiting for the task: (returnval){ [ 767.246605] env[62385]: value = "task-1205623" [ 767.246605] env[62385]: _type = "Task" [ 767.246605] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.258263] env[62385]: DEBUG oslo_vmware.api [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Task: {'id': task-1205623, 'name': Rename_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.276951] env[62385]: DEBUG nova.compute.manager [req-5f747722-0f3a-4784-9695-7389abd8097e req-50639b6e-4321-4ac0-a9a3-673d4a9f01d4 service nova] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Received event network-vif-plugged-bfb1f5cb-27f6-4422-8a40-f8f2d8a6520b {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 767.277195] env[62385]: DEBUG oslo_concurrency.lockutils [req-5f747722-0f3a-4784-9695-7389abd8097e req-50639b6e-4321-4ac0-a9a3-673d4a9f01d4 service nova] Acquiring lock "f8a98710-9d02-4aa3-b6a5-34b7060ad62d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.277407] env[62385]: DEBUG oslo_concurrency.lockutils [req-5f747722-0f3a-4784-9695-7389abd8097e req-50639b6e-4321-4ac0-a9a3-673d4a9f01d4 service nova] Lock "f8a98710-9d02-4aa3-b6a5-34b7060ad62d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.277575] env[62385]: DEBUG oslo_concurrency.lockutils [req-5f747722-0f3a-4784-9695-7389abd8097e req-50639b6e-4321-4ac0-a9a3-673d4a9f01d4 service nova] Lock "f8a98710-9d02-4aa3-b6a5-34b7060ad62d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.277740] env[62385]: DEBUG nova.compute.manager [req-5f747722-0f3a-4784-9695-7389abd8097e req-50639b6e-4321-4ac0-a9a3-673d4a9f01d4 service nova] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] No waiting events found dispatching network-vif-plugged-bfb1f5cb-27f6-4422-8a40-f8f2d8a6520b {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 767.277902] env[62385]: WARNING nova.compute.manager [req-5f747722-0f3a-4784-9695-7389abd8097e req-50639b6e-4321-4ac0-a9a3-673d4a9f01d4 service nova] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Received unexpected event network-vif-plugged-bfb1f5cb-27f6-4422-8a40-f8f2d8a6520b for instance with vm_state building and task_state spawning. [ 767.278096] env[62385]: DEBUG nova.compute.manager [req-5f747722-0f3a-4784-9695-7389abd8097e req-50639b6e-4321-4ac0-a9a3-673d4a9f01d4 service nova] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Received event network-changed-bfb1f5cb-27f6-4422-8a40-f8f2d8a6520b {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 767.278267] env[62385]: DEBUG nova.compute.manager [req-5f747722-0f3a-4784-9695-7389abd8097e req-50639b6e-4321-4ac0-a9a3-673d4a9f01d4 service nova] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Refreshing instance network info cache due to event network-changed-bfb1f5cb-27f6-4422-8a40-f8f2d8a6520b. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 767.278464] env[62385]: DEBUG oslo_concurrency.lockutils [req-5f747722-0f3a-4784-9695-7389abd8097e req-50639b6e-4321-4ac0-a9a3-673d4a9f01d4 service nova] Acquiring lock "refresh_cache-f8a98710-9d02-4aa3-b6a5-34b7060ad62d" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.321362] env[62385]: DEBUG oslo_concurrency.lockutils [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.332748] env[62385]: DEBUG nova.compute.manager [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 767.333623] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4692107-1fe2-44fb-9630-652a289db655 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.530613] env[62385]: DEBUG oslo_vmware.api [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]527573d3-ae2f-4bf4-0f26-849651f10488, 'name': SearchDatastore_Task, 'duration_secs': 0.010031} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.532043] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-70ee3532-5484-45f1-a33d-6cebf1f2e516 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.537370] env[62385]: DEBUG oslo_vmware.api [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Waiting for the task: (returnval){ [ 767.537370] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5270488b-ba4f-21c8-1281-ec5dd20d3283" [ 767.537370] env[62385]: _type = "Task" [ 767.537370] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.547938] env[62385]: DEBUG oslo_vmware.api [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5270488b-ba4f-21c8-1281-ec5dd20d3283, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.712575] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d27da6b-c92c-4845-aab4-0c799b8a76bd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.721402] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a54e8655-a6b3-4f36-b477-74fc2f0eee16 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.752853] env[62385]: DEBUG nova.network.neutron [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 767.758899] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18752e09-e417-4be8-9187-d8ca0a4bc684 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.769437] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e22c0c5-59c8-40cf-a226-9c8636ace383 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.775028] env[62385]: DEBUG oslo_vmware.api [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Task: {'id': task-1205623, 'name': Rename_Task, 'duration_secs': 0.163705} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.775028] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 767.775439] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5d8dd8eb-ada4-4a12-b438-b68b0ff11112 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.786168] env[62385]: DEBUG nova.compute.provider_tree [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 767.800018] env[62385]: DEBUG oslo_vmware.api [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Waiting for the task: (returnval){ [ 767.800018] env[62385]: value = "task-1205624" [ 767.800018] env[62385]: _type = "Task" [ 767.800018] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.809132] env[62385]: DEBUG oslo_vmware.api [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Task: {'id': task-1205624, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.846941] env[62385]: INFO nova.compute.manager [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] instance snapshotting [ 767.850167] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25da1a7f-ddbf-415b-8bb9-40794756e175 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.873025] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d08587c-7e57-4d4b-afd0-f370411cd899 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.037812] env[62385]: DEBUG nova.network.neutron [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Updating instance_info_cache with network_info: [{"id": "bfb1f5cb-27f6-4422-8a40-f8f2d8a6520b", "address": "fa:16:3e:0c:f4:9c", "network": {"id": "ff21dd88-fa02-4bd6-8009-980ec6e72758", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1065789169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aea48968d4464ac0ab01c998c059107d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3ff3baee-99ce-4b51-ae98-efc6163aaab3", "external-id": "nsx-vlan-transportzone-574", "segmentation_id": 574, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbfb1f5cb-27", "ovs_interfaceid": "bfb1f5cb-27f6-4422-8a40-f8f2d8a6520b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.050111] env[62385]: DEBUG oslo_vmware.api [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5270488b-ba4f-21c8-1281-ec5dd20d3283, 'name': SearchDatastore_Task, 'duration_secs': 0.010787} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.050352] env[62385]: DEBUG oslo_concurrency.lockutils [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.050613] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 65bb1d14-84b2-4e75-acdc-dc674a035101/65bb1d14-84b2-4e75-acdc-dc674a035101.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 768.051113] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5699089c-6b02-40eb-b487-706b0f5375d1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.060053] env[62385]: DEBUG oslo_vmware.api [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Waiting for the task: (returnval){ [ 768.060053] env[62385]: value = "task-1205625" [ 768.060053] env[62385]: _type = "Task" [ 768.060053] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.069505] env[62385]: DEBUG oslo_vmware.api [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': task-1205625, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.165094] env[62385]: DEBUG nova.compute.manager [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 768.195757] env[62385]: DEBUG nova.virt.hardware [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 768.195757] env[62385]: DEBUG nova.virt.hardware [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 768.195998] env[62385]: DEBUG nova.virt.hardware [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 768.198097] env[62385]: DEBUG nova.virt.hardware [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 768.198097] env[62385]: DEBUG nova.virt.hardware [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 768.198097] env[62385]: DEBUG nova.virt.hardware [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 768.198097] env[62385]: DEBUG nova.virt.hardware [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 768.198097] env[62385]: DEBUG nova.virt.hardware [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 768.198475] env[62385]: DEBUG nova.virt.hardware [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 768.198475] env[62385]: DEBUG nova.virt.hardware [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 768.198475] env[62385]: DEBUG nova.virt.hardware [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 768.198475] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f607bae2-4846-46bc-a07a-8f36d9e9a272 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.209428] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca155fa8-e2fd-4102-a771-f55532fe3e40 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.290389] env[62385]: DEBUG nova.scheduler.client.report [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 768.316326] env[62385]: DEBUG oslo_vmware.api [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Task: {'id': task-1205624, 'name': PowerOnVM_Task, 'duration_secs': 0.469436} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.316627] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 768.316826] env[62385]: INFO nova.compute.manager [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Took 7.68 seconds to spawn the instance on the hypervisor. [ 768.317060] env[62385]: DEBUG nova.compute.manager [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 768.317960] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34df4e20-3853-41e6-80c6-d5545fd2c992 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.384792] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Creating Snapshot of the VM instance {{(pid=62385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 768.385232] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-02e21a09-8ba2-4097-8395-de8f9caefd04 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.396265] env[62385]: DEBUG oslo_vmware.api [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Waiting for the task: (returnval){ [ 768.396265] env[62385]: value = "task-1205626" [ 768.396265] env[62385]: _type = "Task" [ 768.396265] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.407766] env[62385]: DEBUG oslo_vmware.api [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205626, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.523161] env[62385]: DEBUG nova.compute.manager [req-79be9d90-4c46-4d95-8cac-7391f7945e8d req-7a0b0309-a7e2-48fb-b58c-ecac578a49d7 service nova] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Received event network-vif-plugged-326dbe5f-be23-4610-bc79-079259f3a2c3 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 768.523839] env[62385]: DEBUG oslo_concurrency.lockutils [req-79be9d90-4c46-4d95-8cac-7391f7945e8d req-7a0b0309-a7e2-48fb-b58c-ecac578a49d7 service nova] Acquiring lock "f77dc5ab-7e63-4a1b-9c87-e474a7e28c93-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.523839] env[62385]: DEBUG oslo_concurrency.lockutils [req-79be9d90-4c46-4d95-8cac-7391f7945e8d req-7a0b0309-a7e2-48fb-b58c-ecac578a49d7 service nova] Lock "f77dc5ab-7e63-4a1b-9c87-e474a7e28c93-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.523839] env[62385]: DEBUG oslo_concurrency.lockutils [req-79be9d90-4c46-4d95-8cac-7391f7945e8d req-7a0b0309-a7e2-48fb-b58c-ecac578a49d7 service nova] Lock "f77dc5ab-7e63-4a1b-9c87-e474a7e28c93-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.524087] env[62385]: DEBUG nova.compute.manager [req-79be9d90-4c46-4d95-8cac-7391f7945e8d req-7a0b0309-a7e2-48fb-b58c-ecac578a49d7 service nova] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] No waiting events found dispatching network-vif-plugged-326dbe5f-be23-4610-bc79-079259f3a2c3 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 768.524801] env[62385]: WARNING nova.compute.manager [req-79be9d90-4c46-4d95-8cac-7391f7945e8d req-7a0b0309-a7e2-48fb-b58c-ecac578a49d7 service nova] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Received unexpected event network-vif-plugged-326dbe5f-be23-4610-bc79-079259f3a2c3 for instance with vm_state building and task_state spawning. [ 768.544608] env[62385]: DEBUG oslo_concurrency.lockutils [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Releasing lock "refresh_cache-f8a98710-9d02-4aa3-b6a5-34b7060ad62d" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.544608] env[62385]: DEBUG nova.compute.manager [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Instance network_info: |[{"id": "bfb1f5cb-27f6-4422-8a40-f8f2d8a6520b", "address": "fa:16:3e:0c:f4:9c", "network": {"id": "ff21dd88-fa02-4bd6-8009-980ec6e72758", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1065789169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aea48968d4464ac0ab01c998c059107d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3ff3baee-99ce-4b51-ae98-efc6163aaab3", "external-id": "nsx-vlan-transportzone-574", "segmentation_id": 574, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbfb1f5cb-27", "ovs_interfaceid": "bfb1f5cb-27f6-4422-8a40-f8f2d8a6520b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 768.544906] env[62385]: DEBUG oslo_concurrency.lockutils [req-5f747722-0f3a-4784-9695-7389abd8097e req-50639b6e-4321-4ac0-a9a3-673d4a9f01d4 service nova] Acquired lock "refresh_cache-f8a98710-9d02-4aa3-b6a5-34b7060ad62d" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.544906] env[62385]: DEBUG nova.network.neutron [req-5f747722-0f3a-4784-9695-7389abd8097e req-50639b6e-4321-4ac0-a9a3-673d4a9f01d4 service nova] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Refreshing network info cache for port bfb1f5cb-27f6-4422-8a40-f8f2d8a6520b {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 768.546154] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0c:f4:9c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3ff3baee-99ce-4b51-ae98-efc6163aaab3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bfb1f5cb-27f6-4422-8a40-f8f2d8a6520b', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 768.554098] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Creating folder: Project (aea48968d4464ac0ab01c998c059107d). Parent ref: group-v261107. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 768.557464] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-21672bd8-60fb-4ef2-9f39-056af1116705 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.570990] env[62385]: DEBUG oslo_vmware.api [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': task-1205625, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.572369] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Created folder: Project (aea48968d4464ac0ab01c998c059107d) in parent group-v261107. [ 768.572551] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Creating folder: Instances. Parent ref: group-v261160. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 768.572785] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-74d9c7b2-d161-4a20-8e80-4b29c59e59f2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.584418] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Created folder: Instances in parent group-v261160. [ 768.584678] env[62385]: DEBUG oslo.service.loopingcall [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 768.584878] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 768.585112] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-49be62ff-4bde-47d1-8f8d-53d07383ed0f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.608324] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 768.608324] env[62385]: value = "task-1205629" [ 768.608324] env[62385]: _type = "Task" [ 768.608324] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.617524] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205629, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.647674] env[62385]: DEBUG nova.network.neutron [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Successfully updated port: 326dbe5f-be23-4610-bc79-079259f3a2c3 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 768.798080] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.656s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.798668] env[62385]: DEBUG nova.compute.manager [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 768.802326] env[62385]: DEBUG oslo_concurrency.lockutils [None req-57e5a242-f0aa-45f3-b3cd-46ad0f3354b0 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.535s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.802606] env[62385]: DEBUG nova.objects.instance [None req-57e5a242-f0aa-45f3-b3cd-46ad0f3354b0 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Lazy-loading 'resources' on Instance uuid 3e184125-28af-469b-83cc-4ab8859e0e2f {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 768.846669] env[62385]: INFO nova.compute.manager [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Took 40.81 seconds to build instance. [ 768.906842] env[62385]: DEBUG oslo_vmware.api [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205626, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.931413] env[62385]: DEBUG nova.network.neutron [req-5f747722-0f3a-4784-9695-7389abd8097e req-50639b6e-4321-4ac0-a9a3-673d4a9f01d4 service nova] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Updated VIF entry in instance network info cache for port bfb1f5cb-27f6-4422-8a40-f8f2d8a6520b. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 768.931779] env[62385]: DEBUG nova.network.neutron [req-5f747722-0f3a-4784-9695-7389abd8097e req-50639b6e-4321-4ac0-a9a3-673d4a9f01d4 service nova] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Updating instance_info_cache with network_info: [{"id": "bfb1f5cb-27f6-4422-8a40-f8f2d8a6520b", "address": "fa:16:3e:0c:f4:9c", "network": {"id": "ff21dd88-fa02-4bd6-8009-980ec6e72758", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1065789169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aea48968d4464ac0ab01c998c059107d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3ff3baee-99ce-4b51-ae98-efc6163aaab3", "external-id": "nsx-vlan-transportzone-574", "segmentation_id": 574, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbfb1f5cb-27", "ovs_interfaceid": "bfb1f5cb-27f6-4422-8a40-f8f2d8a6520b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.072449] env[62385]: DEBUG oslo_vmware.api [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': task-1205625, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.519} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.072718] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 65bb1d14-84b2-4e75-acdc-dc674a035101/65bb1d14-84b2-4e75-acdc-dc674a035101.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 769.072921] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 769.073191] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6d513ffc-0b1f-45e0-a248-f8f101c30ebb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.082466] env[62385]: DEBUG oslo_vmware.api [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Waiting for the task: (returnval){ [ 769.082466] env[62385]: value = "task-1205630" [ 769.082466] env[62385]: _type = "Task" [ 769.082466] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.091325] env[62385]: DEBUG oslo_vmware.api [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': task-1205630, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.118415] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205629, 'name': CreateVM_Task} progress is 25%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.150296] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Acquiring lock "refresh_cache-f77dc5ab-7e63-4a1b-9c87-e474a7e28c93" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.150479] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Acquired lock "refresh_cache-f77dc5ab-7e63-4a1b-9c87-e474a7e28c93" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.150666] env[62385]: DEBUG nova.network.neutron [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 769.308032] env[62385]: DEBUG nova.compute.utils [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 769.310640] env[62385]: DEBUG nova.compute.manager [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Not allocating networking since 'none' was specified. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 769.352211] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec5440a6-4a75-48fc-b134-528082dadcb3 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Lock "e73ecefc-fed3-4ccd-88d3-399a4b72bb4b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 132.359s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.412326] env[62385]: DEBUG oslo_vmware.api [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205626, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.436613] env[62385]: DEBUG oslo_concurrency.lockutils [req-5f747722-0f3a-4784-9695-7389abd8097e req-50639b6e-4321-4ac0-a9a3-673d4a9f01d4 service nova] Releasing lock "refresh_cache-f8a98710-9d02-4aa3-b6a5-34b7060ad62d" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.597275] env[62385]: DEBUG oslo_vmware.api [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': task-1205630, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083498} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.600868] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 769.602369] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea0ccf22-ffa5-4edf-abb8-5b53b561f7c2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.627785] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Reconfiguring VM instance instance-00000037 to attach disk [datastore2] 65bb1d14-84b2-4e75-acdc-dc674a035101/65bb1d14-84b2-4e75-acdc-dc674a035101.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 769.633838] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-717cb97e-1946-4f38-a8a1-5513d5c003b1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.657981] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205629, 'name': CreateVM_Task, 'duration_secs': 0.729555} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.659777] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 769.659842] env[62385]: DEBUG oslo_vmware.api [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Waiting for the task: (returnval){ [ 769.659842] env[62385]: value = "task-1205631" [ 769.659842] env[62385]: _type = "Task" [ 769.659842] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.663719] env[62385]: DEBUG oslo_concurrency.lockutils [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.663719] env[62385]: DEBUG oslo_concurrency.lockutils [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.663719] env[62385]: DEBUG oslo_concurrency.lockutils [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 769.664221] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-57d13f4d-144b-4f0a-b391-44809a784525 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.673530] env[62385]: DEBUG nova.compute.manager [req-c8916815-08d9-4c09-a32d-0ca230bfa898 req-4a76a3c3-f268-4766-a4c5-4fdea65ac45c service nova] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Received event network-changed-30e19299-99d0-4c98-8615-78f2f5f76dd5 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 769.673737] env[62385]: DEBUG nova.compute.manager [req-c8916815-08d9-4c09-a32d-0ca230bfa898 req-4a76a3c3-f268-4766-a4c5-4fdea65ac45c service nova] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Refreshing instance network info cache due to event network-changed-30e19299-99d0-4c98-8615-78f2f5f76dd5. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 769.673985] env[62385]: DEBUG oslo_concurrency.lockutils [req-c8916815-08d9-4c09-a32d-0ca230bfa898 req-4a76a3c3-f268-4766-a4c5-4fdea65ac45c service nova] Acquiring lock "refresh_cache-e73ecefc-fed3-4ccd-88d3-399a4b72bb4b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.674144] env[62385]: DEBUG oslo_concurrency.lockutils [req-c8916815-08d9-4c09-a32d-0ca230bfa898 req-4a76a3c3-f268-4766-a4c5-4fdea65ac45c service nova] Acquired lock "refresh_cache-e73ecefc-fed3-4ccd-88d3-399a4b72bb4b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.674310] env[62385]: DEBUG nova.network.neutron [req-c8916815-08d9-4c09-a32d-0ca230bfa898 req-4a76a3c3-f268-4766-a4c5-4fdea65ac45c service nova] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Refreshing network info cache for port 30e19299-99d0-4c98-8615-78f2f5f76dd5 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 769.679801] env[62385]: DEBUG oslo_vmware.api [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 769.679801] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52d2409e-e8a0-1956-1989-52445f91a076" [ 769.679801] env[62385]: _type = "Task" [ 769.679801] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.684155] env[62385]: DEBUG oslo_vmware.api [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': task-1205631, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.701095] env[62385]: DEBUG oslo_vmware.api [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52d2409e-e8a0-1956-1989-52445f91a076, 'name': SearchDatastore_Task, 'duration_secs': 0.015516} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.704140] env[62385]: DEBUG oslo_concurrency.lockutils [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.704403] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 769.704632] env[62385]: DEBUG oslo_concurrency.lockutils [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.704779] env[62385]: DEBUG oslo_concurrency.lockutils [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.704955] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 769.705807] env[62385]: DEBUG nova.network.neutron [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 769.707763] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-26e567c5-ae0a-480e-b8b5-61ee3df12e51 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.723141] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 769.723390] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 769.724311] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d930bf75-d8c8-46c4-9d75-6b2c5146aa48 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.731201] env[62385]: DEBUG oslo_vmware.api [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 769.731201] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]520aeac0-b83b-aa63-2155-ab6713961b9f" [ 769.731201] env[62385]: _type = "Task" [ 769.731201] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.745885] env[62385]: DEBUG oslo_vmware.api [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]520aeac0-b83b-aa63-2155-ab6713961b9f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.775518] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-130d992c-2a4b-405c-9caf-ac4222e6fdd2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.787795] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b5642d6-c2f2-4c8b-8232-c2d78d444bbf {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.820105] env[62385]: DEBUG nova.compute.manager [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 769.825811] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5264457-7382-4e59-99c0-82a8609a2f41 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.834790] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e206c57-b3aa-456a-b8db-11608d31c626 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.850965] env[62385]: DEBUG nova.compute.provider_tree [None req-57e5a242-f0aa-45f3-b3cd-46ad0f3354b0 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 769.858212] env[62385]: DEBUG nova.compute.manager [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] [instance: 2b744879-01d2-440e-8adf-58438b922888] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 769.907934] env[62385]: DEBUG oslo_vmware.api [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205626, 'name': CreateSnapshot_Task, 'duration_secs': 1.466962} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.908660] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Created Snapshot of the VM instance {{(pid=62385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 769.909497] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abc6c6f4-20b5-47af-9b72-01ddc71c445e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.970964] env[62385]: DEBUG nova.network.neutron [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Updating instance_info_cache with network_info: [{"id": "326dbe5f-be23-4610-bc79-079259f3a2c3", "address": "fa:16:3e:0a:bf:43", "network": {"id": "63f9709f-522a-4b61-b021-51efd6f18635", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1946207557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06ca138339954db79c9bd6af2bcac5ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b0fa7a2-ebd9-4788-8904-7bf250ce466c", "external-id": "nsx-vlan-transportzone-669", "segmentation_id": 669, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap326dbe5f-be", "ovs_interfaceid": "326dbe5f-be23-4610-bc79-079259f3a2c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.175347] env[62385]: DEBUG oslo_vmware.api [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': task-1205631, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.242265] env[62385]: DEBUG oslo_vmware.api [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]520aeac0-b83b-aa63-2155-ab6713961b9f, 'name': SearchDatastore_Task, 'duration_secs': 0.017875} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.243131] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3bda276-15a7-4249-9765-f8b38cf0c793 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.248824] env[62385]: DEBUG oslo_vmware.api [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 770.248824] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5231f5a3-11eb-498d-6360-52a65ad0c5fc" [ 770.248824] env[62385]: _type = "Task" [ 770.248824] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.256926] env[62385]: DEBUG oslo_vmware.api [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5231f5a3-11eb-498d-6360-52a65ad0c5fc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.356813] env[62385]: DEBUG nova.scheduler.client.report [None req-57e5a242-f0aa-45f3-b3cd-46ad0f3354b0 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 770.360093] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f1ed29c2-f0f6-4349-82f1-05918a2b3238 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Acquiring lock "e73ecefc-fed3-4ccd-88d3-399a4b72bb4b" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.360325] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f1ed29c2-f0f6-4349-82f1-05918a2b3238 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Lock "e73ecefc-fed3-4ccd-88d3-399a4b72bb4b" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.360611] env[62385]: INFO nova.compute.manager [None req-f1ed29c2-f0f6-4349-82f1-05918a2b3238 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Rebooting instance [ 770.384742] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.415782] env[62385]: DEBUG nova.network.neutron [req-c8916815-08d9-4c09-a32d-0ca230bfa898 req-4a76a3c3-f268-4766-a4c5-4fdea65ac45c service nova] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Updated VIF entry in instance network info cache for port 30e19299-99d0-4c98-8615-78f2f5f76dd5. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 770.416230] env[62385]: DEBUG nova.network.neutron [req-c8916815-08d9-4c09-a32d-0ca230bfa898 req-4a76a3c3-f268-4766-a4c5-4fdea65ac45c service nova] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Updating instance_info_cache with network_info: [{"id": "30e19299-99d0-4c98-8615-78f2f5f76dd5", "address": "fa:16:3e:a9:58:3d", "network": {"id": "63367e0f-9855-478b-a546-5f0cf8190519", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-21852824-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ebc1c1ed0bb404b89e59325535edecd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "778b9a40-d603-4765-ac88-bd6d42c457a2", "external-id": "nsx-vlan-transportzone-114", "segmentation_id": 114, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30e19299-99", "ovs_interfaceid": "30e19299-99d0-4c98-8615-78f2f5f76dd5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.429956] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Creating linked-clone VM from snapshot {{(pid=62385) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 770.430668] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-658ae373-a09a-443f-a00d-48780d4454b2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.440884] env[62385]: DEBUG oslo_vmware.api [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Waiting for the task: (returnval){ [ 770.440884] env[62385]: value = "task-1205632" [ 770.440884] env[62385]: _type = "Task" [ 770.440884] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.449161] env[62385]: DEBUG oslo_vmware.api [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205632, 'name': CloneVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.473385] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Releasing lock "refresh_cache-f77dc5ab-7e63-4a1b-9c87-e474a7e28c93" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.473720] env[62385]: DEBUG nova.compute.manager [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Instance network_info: |[{"id": "326dbe5f-be23-4610-bc79-079259f3a2c3", "address": "fa:16:3e:0a:bf:43", "network": {"id": "63f9709f-522a-4b61-b021-51efd6f18635", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1946207557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06ca138339954db79c9bd6af2bcac5ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b0fa7a2-ebd9-4788-8904-7bf250ce466c", "external-id": "nsx-vlan-transportzone-669", "segmentation_id": 669, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap326dbe5f-be", "ovs_interfaceid": "326dbe5f-be23-4610-bc79-079259f3a2c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 770.474130] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0a:bf:43', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5b0fa7a2-ebd9-4788-8904-7bf250ce466c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '326dbe5f-be23-4610-bc79-079259f3a2c3', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 770.482386] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Creating folder: Project (06ca138339954db79c9bd6af2bcac5ff). Parent ref: group-v261107. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 770.482386] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3628dbe9-3f96-444b-a0bf-dfa3a1157261 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.502045] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Created folder: Project (06ca138339954db79c9bd6af2bcac5ff) in parent group-v261107. [ 770.502291] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Creating folder: Instances. Parent ref: group-v261165. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 770.502625] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8120f226-ef8e-4d0e-8c40-612d4af63e22 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.515311] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Created folder: Instances in parent group-v261165. [ 770.515645] env[62385]: DEBUG oslo.service.loopingcall [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 770.515901] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 770.516323] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f1c9ce3a-e5d8-4cab-a505-1aa3af4ee27b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.537788] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 770.537788] env[62385]: value = "task-1205635" [ 770.537788] env[62385]: _type = "Task" [ 770.537788] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.545812] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205635, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.585973] env[62385]: DEBUG nova.compute.manager [req-59569088-85a6-4370-bff8-993a6859ea42 req-fd0245ec-baf5-448e-b9ab-da490e687141 service nova] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Received event network-changed-326dbe5f-be23-4610-bc79-079259f3a2c3 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 770.586129] env[62385]: DEBUG nova.compute.manager [req-59569088-85a6-4370-bff8-993a6859ea42 req-fd0245ec-baf5-448e-b9ab-da490e687141 service nova] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Refreshing instance network info cache due to event network-changed-326dbe5f-be23-4610-bc79-079259f3a2c3. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 770.586769] env[62385]: DEBUG oslo_concurrency.lockutils [req-59569088-85a6-4370-bff8-993a6859ea42 req-fd0245ec-baf5-448e-b9ab-da490e687141 service nova] Acquiring lock "refresh_cache-f77dc5ab-7e63-4a1b-9c87-e474a7e28c93" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 770.586769] env[62385]: DEBUG oslo_concurrency.lockutils [req-59569088-85a6-4370-bff8-993a6859ea42 req-fd0245ec-baf5-448e-b9ab-da490e687141 service nova] Acquired lock "refresh_cache-f77dc5ab-7e63-4a1b-9c87-e474a7e28c93" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.586769] env[62385]: DEBUG nova.network.neutron [req-59569088-85a6-4370-bff8-993a6859ea42 req-fd0245ec-baf5-448e-b9ab-da490e687141 service nova] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Refreshing network info cache for port 326dbe5f-be23-4610-bc79-079259f3a2c3 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 770.675408] env[62385]: DEBUG oslo_vmware.api [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': task-1205631, 'name': ReconfigVM_Task, 'duration_secs': 0.657567} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.675711] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Reconfigured VM instance instance-00000037 to attach disk [datastore2] 65bb1d14-84b2-4e75-acdc-dc674a035101/65bb1d14-84b2-4e75-acdc-dc674a035101.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 770.676433] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-93424004-7aac-4205-8d78-c76503f53cd3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.684573] env[62385]: DEBUG oslo_vmware.api [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Waiting for the task: (returnval){ [ 770.684573] env[62385]: value = "task-1205636" [ 770.684573] env[62385]: _type = "Task" [ 770.684573] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.694224] env[62385]: DEBUG oslo_vmware.api [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': task-1205636, 'name': Rename_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.760834] env[62385]: DEBUG oslo_vmware.api [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5231f5a3-11eb-498d-6360-52a65ad0c5fc, 'name': SearchDatastore_Task, 'duration_secs': 0.013992} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.761106] env[62385]: DEBUG oslo_concurrency.lockutils [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.761379] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] f8a98710-9d02-4aa3-b6a5-34b7060ad62d/f8a98710-9d02-4aa3-b6a5-34b7060ad62d.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 770.761643] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1d7594ed-933b-4f58-a7b3-c1578626cdae {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.770088] env[62385]: DEBUG oslo_vmware.api [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 770.770088] env[62385]: value = "task-1205637" [ 770.770088] env[62385]: _type = "Task" [ 770.770088] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.779970] env[62385]: DEBUG oslo_vmware.api [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205637, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.835008] env[62385]: DEBUG nova.compute.manager [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 770.865195] env[62385]: DEBUG nova.virt.hardware [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 770.865548] env[62385]: DEBUG nova.virt.hardware [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 770.865774] env[62385]: DEBUG nova.virt.hardware [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 770.866072] env[62385]: DEBUG nova.virt.hardware [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 770.866301] env[62385]: DEBUG nova.virt.hardware [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 770.866550] env[62385]: DEBUG nova.virt.hardware [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 770.866851] env[62385]: DEBUG nova.virt.hardware [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 770.867149] env[62385]: DEBUG nova.virt.hardware [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 770.867428] env[62385]: DEBUG nova.virt.hardware [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 770.867684] env[62385]: DEBUG nova.virt.hardware [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 770.867944] env[62385]: DEBUG nova.virt.hardware [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 770.868948] env[62385]: DEBUG oslo_concurrency.lockutils [None req-57e5a242-f0aa-45f3-b3cd-46ad0f3354b0 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.067s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.872400] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-843c5e4f-0005-454f-93d6-c704fdbbb8a7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.876263] env[62385]: DEBUG oslo_concurrency.lockutils [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.342s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.878546] env[62385]: INFO nova.compute.claims [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 770.890805] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb7072e1-b0d8-4f80-a55f-93e573195112 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.908360] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Instance VIF info [] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 770.915273] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Creating folder: Project (d552200e0d6149f2bf4a4f8fec458a98). Parent ref: group-v261107. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 770.916703] env[62385]: INFO nova.scheduler.client.report [None req-57e5a242-f0aa-45f3-b3cd-46ad0f3354b0 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Deleted allocations for instance 3e184125-28af-469b-83cc-4ab8859e0e2f [ 770.918684] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f1ed29c2-f0f6-4349-82f1-05918a2b3238 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Acquiring lock "refresh_cache-e73ecefc-fed3-4ccd-88d3-399a4b72bb4b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 770.918959] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bb243e65-a1c2-49a4-965e-17bc90246548 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.921363] env[62385]: DEBUG oslo_concurrency.lockutils [req-c8916815-08d9-4c09-a32d-0ca230bfa898 req-4a76a3c3-f268-4766-a4c5-4fdea65ac45c service nova] Releasing lock "refresh_cache-e73ecefc-fed3-4ccd-88d3-399a4b72bb4b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.924912] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f1ed29c2-f0f6-4349-82f1-05918a2b3238 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Acquired lock "refresh_cache-e73ecefc-fed3-4ccd-88d3-399a4b72bb4b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.925790] env[62385]: DEBUG nova.network.neutron [None req-f1ed29c2-f0f6-4349-82f1-05918a2b3238 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 770.941486] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Created folder: Project (d552200e0d6149f2bf4a4f8fec458a98) in parent group-v261107. [ 770.941699] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Creating folder: Instances. Parent ref: group-v261168. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 770.942039] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-94feb0a0-9e24-430d-8f44-1114ca888ca3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.954196] env[62385]: DEBUG oslo_vmware.api [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205632, 'name': CloneVM_Task} progress is 94%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.960240] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Created folder: Instances in parent group-v261168. [ 770.960586] env[62385]: DEBUG oslo.service.loopingcall [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 770.960775] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 770.961029] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ca2731cd-d313-4ce9-8287-f5d1bde4c34f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.980824] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 770.980824] env[62385]: value = "task-1205640" [ 770.980824] env[62385]: _type = "Task" [ 770.980824] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.993578] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205640, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.059622] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205635, 'name': CreateVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.197392] env[62385]: DEBUG oslo_vmware.api [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': task-1205636, 'name': Rename_Task, 'duration_secs': 0.386203} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.197940] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 771.198378] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-13b683be-b6f3-45a0-8fbd-fcb290a8004f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.206663] env[62385]: DEBUG oslo_vmware.api [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Waiting for the task: (returnval){ [ 771.206663] env[62385]: value = "task-1205641" [ 771.206663] env[62385]: _type = "Task" [ 771.206663] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.217245] env[62385]: DEBUG oslo_vmware.api [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': task-1205641, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.286814] env[62385]: DEBUG oslo_vmware.api [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205637, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.497065} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.287345] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] f8a98710-9d02-4aa3-b6a5-34b7060ad62d/f8a98710-9d02-4aa3-b6a5-34b7060ad62d.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 771.288575] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 771.288575] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3a264bb7-c100-4783-bef7-021dc6d10d58 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.301762] env[62385]: DEBUG oslo_vmware.api [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 771.301762] env[62385]: value = "task-1205642" [ 771.301762] env[62385]: _type = "Task" [ 771.301762] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.313537] env[62385]: DEBUG oslo_vmware.api [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205642, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.431712] env[62385]: DEBUG oslo_concurrency.lockutils [None req-57e5a242-f0aa-45f3-b3cd-46ad0f3354b0 tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Lock "3e184125-28af-469b-83cc-4ab8859e0e2f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.205s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.453081] env[62385]: DEBUG oslo_vmware.api [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205632, 'name': CloneVM_Task} progress is 94%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.492236] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205640, 'name': CreateVM_Task, 'duration_secs': 0.413371} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.492414] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 771.492845] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.492998] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.493335] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 771.493581] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-821bd632-4f25-4be5-b8d3-7ff2a2fff31a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.498680] env[62385]: DEBUG oslo_vmware.api [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Waiting for the task: (returnval){ [ 771.498680] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5289e0ca-1cb1-3496-01af-c2684e1bb228" [ 771.498680] env[62385]: _type = "Task" [ 771.498680] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.506917] env[62385]: DEBUG oslo_vmware.api [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5289e0ca-1cb1-3496-01af-c2684e1bb228, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.517126] env[62385]: DEBUG nova.network.neutron [req-59569088-85a6-4370-bff8-993a6859ea42 req-fd0245ec-baf5-448e-b9ab-da490e687141 service nova] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Updated VIF entry in instance network info cache for port 326dbe5f-be23-4610-bc79-079259f3a2c3. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 771.517126] env[62385]: DEBUG nova.network.neutron [req-59569088-85a6-4370-bff8-993a6859ea42 req-fd0245ec-baf5-448e-b9ab-da490e687141 service nova] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Updating instance_info_cache with network_info: [{"id": "326dbe5f-be23-4610-bc79-079259f3a2c3", "address": "fa:16:3e:0a:bf:43", "network": {"id": "63f9709f-522a-4b61-b021-51efd6f18635", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1946207557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06ca138339954db79c9bd6af2bcac5ff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b0fa7a2-ebd9-4788-8904-7bf250ce466c", "external-id": "nsx-vlan-transportzone-669", "segmentation_id": 669, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap326dbe5f-be", "ovs_interfaceid": "326dbe5f-be23-4610-bc79-079259f3a2c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.549674] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205635, 'name': CreateVM_Task, 'duration_secs': 0.713136} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.549841] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 771.550524] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.550699] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.551021] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 771.551282] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2ed71a1-728c-41c7-bdc5-e7bde2d11de9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.558404] env[62385]: DEBUG oslo_vmware.api [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Waiting for the task: (returnval){ [ 771.558404] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52fc8396-7f4f-f754-ed1e-2ef3af6c90c6" [ 771.558404] env[62385]: _type = "Task" [ 771.558404] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.566963] env[62385]: DEBUG oslo_vmware.api [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52fc8396-7f4f-f754-ed1e-2ef3af6c90c6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.647951] env[62385]: DEBUG nova.network.neutron [None req-f1ed29c2-f0f6-4349-82f1-05918a2b3238 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Updating instance_info_cache with network_info: [{"id": "30e19299-99d0-4c98-8615-78f2f5f76dd5", "address": "fa:16:3e:a9:58:3d", "network": {"id": "63367e0f-9855-478b-a546-5f0cf8190519", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-21852824-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ebc1c1ed0bb404b89e59325535edecd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "778b9a40-d603-4765-ac88-bd6d42c457a2", "external-id": "nsx-vlan-transportzone-114", "segmentation_id": 114, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30e19299-99", "ovs_interfaceid": "30e19299-99d0-4c98-8615-78f2f5f76dd5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.716660] env[62385]: DEBUG oslo_vmware.api [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': task-1205641, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.813783] env[62385]: DEBUG oslo_vmware.api [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205642, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074702} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.813986] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 771.814835] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef499ae5-8f5a-477a-a9b0-7dc66b677a4e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.837263] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Reconfiguring VM instance instance-00000038 to attach disk [datastore2] f8a98710-9d02-4aa3-b6a5-34b7060ad62d/f8a98710-9d02-4aa3-b6a5-34b7060ad62d.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 771.837848] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4bab61d3-eb7c-4abf-b03a-1ad7297ddda9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.857149] env[62385]: DEBUG oslo_vmware.api [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 771.857149] env[62385]: value = "task-1205643" [ 771.857149] env[62385]: _type = "Task" [ 771.857149] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.868089] env[62385]: DEBUG oslo_vmware.api [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205643, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.955467] env[62385]: DEBUG oslo_vmware.api [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205632, 'name': CloneVM_Task, 'duration_secs': 1.304826} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.955467] env[62385]: INFO nova.virt.vmwareapi.vmops [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Created linked-clone VM from snapshot [ 771.956288] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9de6635-253b-46d9-8471-a42d21dde990 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.963227] env[62385]: DEBUG nova.virt.vmwareapi.images [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Uploading image e1482d72-7837-486f-b780-15965cbc5977 {{(pid=62385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 771.989882] env[62385]: DEBUG oslo_vmware.rw_handles [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 771.989882] env[62385]: value = "vm-261164" [ 771.989882] env[62385]: _type = "VirtualMachine" [ 771.989882] env[62385]: }. {{(pid=62385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 771.990996] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-43fb0609-5611-4fb9-ab88-649bc9a2a94e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.997442] env[62385]: DEBUG oslo_vmware.rw_handles [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Lease: (returnval){ [ 771.997442] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5288bb9b-146b-09e5-bbc3-6918db846fae" [ 771.997442] env[62385]: _type = "HttpNfcLease" [ 771.997442] env[62385]: } obtained for exporting VM: (result){ [ 771.997442] env[62385]: value = "vm-261164" [ 771.997442] env[62385]: _type = "VirtualMachine" [ 771.997442] env[62385]: }. {{(pid=62385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 771.998232] env[62385]: DEBUG oslo_vmware.api [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Waiting for the lease: (returnval){ [ 771.998232] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5288bb9b-146b-09e5-bbc3-6918db846fae" [ 771.998232] env[62385]: _type = "HttpNfcLease" [ 771.998232] env[62385]: } to be ready. {{(pid=62385) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 772.008499] env[62385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 772.008499] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5288bb9b-146b-09e5-bbc3-6918db846fae" [ 772.008499] env[62385]: _type = "HttpNfcLease" [ 772.008499] env[62385]: } is initializing. {{(pid=62385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 772.011824] env[62385]: DEBUG oslo_vmware.api [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5289e0ca-1cb1-3496-01af-c2684e1bb228, 'name': SearchDatastore_Task, 'duration_secs': 0.010885} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.014165] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 772.014425] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 772.014651] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.014796] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.014975] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 772.016020] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-df3f08fc-95f7-4eaa-9ef7-2007667a051c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.019691] env[62385]: DEBUG oslo_concurrency.lockutils [req-59569088-85a6-4370-bff8-993a6859ea42 req-fd0245ec-baf5-448e-b9ab-da490e687141 service nova] Releasing lock "refresh_cache-f77dc5ab-7e63-4a1b-9c87-e474a7e28c93" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 772.031983] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 772.032231] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 772.033083] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-606da9a0-5f4c-4769-b12b-45b7a2ad46f4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.040625] env[62385]: DEBUG oslo_vmware.api [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Waiting for the task: (returnval){ [ 772.040625] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52db76ce-2daf-dcc8-9aea-eda4ee8e9be1" [ 772.040625] env[62385]: _type = "Task" [ 772.040625] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.047635] env[62385]: DEBUG oslo_vmware.api [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52db76ce-2daf-dcc8-9aea-eda4ee8e9be1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.069065] env[62385]: DEBUG oslo_vmware.api [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52fc8396-7f4f-f754-ed1e-2ef3af6c90c6, 'name': SearchDatastore_Task, 'duration_secs': 0.010587} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.069363] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 772.069594] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 772.069820] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.069962] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.070148] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 772.070395] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5f1f6c65-8063-468f-8035-38048b37a6cd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.081035] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 772.081222] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 772.083870] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-263c165a-1616-4d59-8212-98fe90929981 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.088551] env[62385]: DEBUG oslo_vmware.api [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Waiting for the task: (returnval){ [ 772.088551] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52b8dcb1-105a-8870-3594-45557c7d4f17" [ 772.088551] env[62385]: _type = "Task" [ 772.088551] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.095500] env[62385]: DEBUG oslo_vmware.api [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52b8dcb1-105a-8870-3594-45557c7d4f17, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.152296] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f1ed29c2-f0f6-4349-82f1-05918a2b3238 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Releasing lock "refresh_cache-e73ecefc-fed3-4ccd-88d3-399a4b72bb4b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 772.156136] env[62385]: DEBUG nova.compute.manager [None req-f1ed29c2-f0f6-4349-82f1-05918a2b3238 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 772.156136] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24ecc367-adf1-454e-9c82-53b11c692cb8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.220204] env[62385]: DEBUG oslo_vmware.api [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': task-1205641, 'name': PowerOnVM_Task, 'duration_secs': 0.594961} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.220512] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 772.220717] env[62385]: INFO nova.compute.manager [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Took 9.07 seconds to spawn the instance on the hypervisor. [ 772.220899] env[62385]: DEBUG nova.compute.manager [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 772.221685] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b21c230a-99d8-490e-b425-348a05b78ba7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.250743] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e2ff260-8a5f-4a99-b3e4-89ba06163a05 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.257744] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdd06845-658d-4161-b6a6-958c9b1d3c45 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.287969] env[62385]: DEBUG oslo_concurrency.lockutils [None req-181b8245-1e74-4984-9c8b-2576aab7b28f tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Acquiring lock "4a4b254c-46a3-4db6-9c0f-f476ae8f2b69" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.288262] env[62385]: DEBUG oslo_concurrency.lockutils [None req-181b8245-1e74-4984-9c8b-2576aab7b28f tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Lock "4a4b254c-46a3-4db6-9c0f-f476ae8f2b69" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.288512] env[62385]: DEBUG oslo_concurrency.lockutils [None req-181b8245-1e74-4984-9c8b-2576aab7b28f tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Acquiring lock "4a4b254c-46a3-4db6-9c0f-f476ae8f2b69-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.288706] env[62385]: DEBUG oslo_concurrency.lockutils [None req-181b8245-1e74-4984-9c8b-2576aab7b28f tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Lock "4a4b254c-46a3-4db6-9c0f-f476ae8f2b69-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.288874] env[62385]: DEBUG oslo_concurrency.lockutils [None req-181b8245-1e74-4984-9c8b-2576aab7b28f tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Lock "4a4b254c-46a3-4db6-9c0f-f476ae8f2b69-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.291907] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bfa8e87-f954-4e1d-8516-43a7714e288c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.294603] env[62385]: INFO nova.compute.manager [None req-181b8245-1e74-4984-9c8b-2576aab7b28f tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Terminating instance [ 772.297918] env[62385]: DEBUG nova.compute.manager [None req-181b8245-1e74-4984-9c8b-2576aab7b28f tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 772.298168] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-181b8245-1e74-4984-9c8b-2576aab7b28f tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 772.299092] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2633c88-4fab-4b7e-aeac-510d85f5d0c6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.305161] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2246a1b9-c522-4238-8510-69b2ffc970d2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.310835] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-181b8245-1e74-4984-9c8b-2576aab7b28f tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 772.311355] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f35d1611-bb31-4ae7-bfc3-378d4fd0654e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.320442] env[62385]: DEBUG nova.compute.provider_tree [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 772.325871] env[62385]: DEBUG oslo_vmware.api [None req-181b8245-1e74-4984-9c8b-2576aab7b28f tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Waiting for the task: (returnval){ [ 772.325871] env[62385]: value = "task-1205645" [ 772.325871] env[62385]: _type = "Task" [ 772.325871] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.335452] env[62385]: DEBUG oslo_vmware.api [None req-181b8245-1e74-4984-9c8b-2576aab7b28f tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205645, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.366958] env[62385]: DEBUG oslo_vmware.api [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205643, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.506558] env[62385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 772.506558] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5288bb9b-146b-09e5-bbc3-6918db846fae" [ 772.506558] env[62385]: _type = "HttpNfcLease" [ 772.506558] env[62385]: } is ready. {{(pid=62385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 772.506887] env[62385]: DEBUG oslo_vmware.rw_handles [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 772.506887] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5288bb9b-146b-09e5-bbc3-6918db846fae" [ 772.506887] env[62385]: _type = "HttpNfcLease" [ 772.506887] env[62385]: }. {{(pid=62385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 772.507663] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f09505d-29d0-47b2-8223-4f06afae66c4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.515550] env[62385]: DEBUG oslo_vmware.rw_handles [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52bc6476-0b00-6dd2-14c5-9b96ff5f7900/disk-0.vmdk from lease info. {{(pid=62385) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 772.515681] env[62385]: DEBUG oslo_vmware.rw_handles [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52bc6476-0b00-6dd2-14c5-9b96ff5f7900/disk-0.vmdk for reading. {{(pid=62385) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 772.584701] env[62385]: DEBUG oslo_vmware.api [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52db76ce-2daf-dcc8-9aea-eda4ee8e9be1, 'name': SearchDatastore_Task, 'duration_secs': 0.013413} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.585783] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c174667-c9c6-4df8-bb6e-feb607480642 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.593463] env[62385]: DEBUG oslo_vmware.api [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Waiting for the task: (returnval){ [ 772.593463] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a8e07c-2a2c-dcef-7a65-b35a864f8a5e" [ 772.593463] env[62385]: _type = "Task" [ 772.593463] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.599826] env[62385]: DEBUG oslo_vmware.api [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52b8dcb1-105a-8870-3594-45557c7d4f17, 'name': SearchDatastore_Task, 'duration_secs': 0.045574} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.600855] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d97a594-ef69-477b-831e-7d7eaa735a07 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.605466] env[62385]: DEBUG oslo_vmware.api [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a8e07c-2a2c-dcef-7a65-b35a864f8a5e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.607967] env[62385]: DEBUG oslo_vmware.api [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Waiting for the task: (returnval){ [ 772.607967] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52cf0474-f0cc-17be-7975-c0852ce6c694" [ 772.607967] env[62385]: _type = "Task" [ 772.607967] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.614523] env[62385]: DEBUG oslo_vmware.api [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52cf0474-f0cc-17be-7975-c0852ce6c694, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.687415] env[62385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-19f51f8f-1f25-4ee4-bf8b-6a8a088a1a91 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.745016] env[62385]: INFO nova.compute.manager [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Took 38.07 seconds to build instance. [ 772.827074] env[62385]: DEBUG nova.scheduler.client.report [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 772.841586] env[62385]: DEBUG oslo_vmware.api [None req-181b8245-1e74-4984-9c8b-2576aab7b28f tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205645, 'name': PowerOffVM_Task, 'duration_secs': 0.501207} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.841586] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-181b8245-1e74-4984-9c8b-2576aab7b28f tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 772.841815] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-181b8245-1e74-4984-9c8b-2576aab7b28f tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 772.841991] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9af17953-9ae1-4a02-bd21-34ab2bbcc188 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.868280] env[62385]: DEBUG oslo_vmware.api [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205643, 'name': ReconfigVM_Task, 'duration_secs': 0.740376} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.868280] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Reconfigured VM instance instance-00000038 to attach disk [datastore2] f8a98710-9d02-4aa3-b6a5-34b7060ad62d/f8a98710-9d02-4aa3-b6a5-34b7060ad62d.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 772.868493] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a5288dc5-1e96-467d-8fca-96be79b9d0a0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.875400] env[62385]: DEBUG oslo_vmware.api [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 772.875400] env[62385]: value = "task-1205647" [ 772.875400] env[62385]: _type = "Task" [ 772.875400] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.886966] env[62385]: DEBUG oslo_vmware.api [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205647, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.923710] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-181b8245-1e74-4984-9c8b-2576aab7b28f tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 772.923943] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-181b8245-1e74-4984-9c8b-2576aab7b28f tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 772.924189] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-181b8245-1e74-4984-9c8b-2576aab7b28f tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Deleting the datastore file [datastore1] 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 772.924490] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c5f95d32-3973-4aef-a155-f3da54c60e70 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.935235] env[62385]: DEBUG oslo_vmware.api [None req-181b8245-1e74-4984-9c8b-2576aab7b28f tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Waiting for the task: (returnval){ [ 772.935235] env[62385]: value = "task-1205648" [ 772.935235] env[62385]: _type = "Task" [ 772.935235] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.942480] env[62385]: DEBUG oslo_vmware.api [None req-181b8245-1e74-4984-9c8b-2576aab7b28f tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205648, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.103962] env[62385]: DEBUG oslo_vmware.api [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a8e07c-2a2c-dcef-7a65-b35a864f8a5e, 'name': SearchDatastore_Task, 'duration_secs': 0.024138} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.104311] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.104790] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7/8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 773.105814] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6f8f1722-939c-400b-b7e3-a3ef8537ff49 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.123025] env[62385]: DEBUG oslo_vmware.api [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Waiting for the task: (returnval){ [ 773.123025] env[62385]: value = "task-1205649" [ 773.123025] env[62385]: _type = "Task" [ 773.123025] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.130044] env[62385]: DEBUG oslo_vmware.api [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52cf0474-f0cc-17be-7975-c0852ce6c694, 'name': SearchDatastore_Task, 'duration_secs': 0.043077} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.135550] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.135969] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] f77dc5ab-7e63-4a1b-9c87-e474a7e28c93/f77dc5ab-7e63-4a1b-9c87-e474a7e28c93.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 773.136507] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9f372d45-12a5-4c60-a162-184f56cc3d41 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.144701] env[62385]: DEBUG oslo_vmware.api [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': task-1205649, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.146863] env[62385]: DEBUG oslo_vmware.api [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Waiting for the task: (returnval){ [ 773.146863] env[62385]: value = "task-1205650" [ 773.146863] env[62385]: _type = "Task" [ 773.146863] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.155981] env[62385]: DEBUG oslo_vmware.api [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Task: {'id': task-1205650, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.173231] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad6c53ef-c39a-4d79-a7d4-160e7cca77f6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.182080] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-f1ed29c2-f0f6-4349-82f1-05918a2b3238 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Doing hard reboot of VM {{(pid=62385) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1064}} [ 773.182465] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-a8930dc8-3465-4907-b961-70bcbb07a6c6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.189373] env[62385]: DEBUG oslo_vmware.api [None req-f1ed29c2-f0f6-4349-82f1-05918a2b3238 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Waiting for the task: (returnval){ [ 773.189373] env[62385]: value = "task-1205651" [ 773.189373] env[62385]: _type = "Task" [ 773.189373] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.198961] env[62385]: DEBUG oslo_vmware.api [None req-f1ed29c2-f0f6-4349-82f1-05918a2b3238 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Task: {'id': task-1205651, 'name': ResetVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.246913] env[62385]: DEBUG oslo_concurrency.lockutils [None req-69f2415c-8ab5-4ef2-9663-fb27a510599e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Lock "65bb1d14-84b2-4e75-acdc-dc674a035101" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 130.471s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.334122] env[62385]: DEBUG oslo_concurrency.lockutils [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.458s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.335491] env[62385]: DEBUG nova.compute.manager [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 773.339224] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8b788b1c-9691-4fa5-a523-a8b477c9025c tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.487s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 773.342152] env[62385]: DEBUG nova.objects.instance [None req-8b788b1c-9691-4fa5-a523-a8b477c9025c tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Lazy-loading 'resources' on Instance uuid 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 773.392884] env[62385]: DEBUG oslo_vmware.api [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205647, 'name': Rename_Task, 'duration_secs': 0.153984} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.394370] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 773.394921] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8a269dea-1c95-48e7-beb4-c1650dab71d0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.407119] env[62385]: DEBUG oslo_vmware.api [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 773.407119] env[62385]: value = "task-1205652" [ 773.407119] env[62385]: _type = "Task" [ 773.407119] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.416609] env[62385]: INFO nova.compute.manager [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Rescuing [ 773.417146] env[62385]: DEBUG oslo_concurrency.lockutils [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Acquiring lock "refresh_cache-65bb1d14-84b2-4e75-acdc-dc674a035101" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.417477] env[62385]: DEBUG oslo_concurrency.lockutils [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Acquired lock "refresh_cache-65bb1d14-84b2-4e75-acdc-dc674a035101" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.417820] env[62385]: DEBUG nova.network.neutron [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 773.425509] env[62385]: DEBUG oslo_vmware.api [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205652, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.445114] env[62385]: DEBUG oslo_vmware.api [None req-181b8245-1e74-4984-9c8b-2576aab7b28f tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205648, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.180912} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.445273] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-181b8245-1e74-4984-9c8b-2576aab7b28f tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 773.445593] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-181b8245-1e74-4984-9c8b-2576aab7b28f tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 773.445886] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-181b8245-1e74-4984-9c8b-2576aab7b28f tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 773.446302] env[62385]: INFO nova.compute.manager [None req-181b8245-1e74-4984-9c8b-2576aab7b28f tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Took 1.15 seconds to destroy the instance on the hypervisor. [ 773.446687] env[62385]: DEBUG oslo.service.loopingcall [None req-181b8245-1e74-4984-9c8b-2576aab7b28f tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 773.446994] env[62385]: DEBUG nova.compute.manager [-] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 773.447157] env[62385]: DEBUG nova.network.neutron [-] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 773.643883] env[62385]: DEBUG oslo_vmware.api [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': task-1205649, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.662843] env[62385]: DEBUG oslo_vmware.api [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Task: {'id': task-1205650, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.701813] env[62385]: DEBUG oslo_vmware.api [None req-f1ed29c2-f0f6-4349-82f1-05918a2b3238 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Task: {'id': task-1205651, 'name': ResetVM_Task, 'duration_secs': 0.09861} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.703623] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-f1ed29c2-f0f6-4349-82f1-05918a2b3238 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Did hard reboot of VM {{(pid=62385) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1068}} [ 773.703623] env[62385]: DEBUG nova.compute.manager [None req-f1ed29c2-f0f6-4349-82f1-05918a2b3238 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 773.703623] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6efcb817-6b5d-4253-8a2e-083d017ab611 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.753115] env[62385]: DEBUG nova.compute.manager [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 773.843605] env[62385]: DEBUG nova.compute.utils [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 773.849561] env[62385]: DEBUG nova.compute.manager [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 773.849672] env[62385]: DEBUG nova.network.neutron [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 773.896109] env[62385]: DEBUG nova.policy [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5040b20c848c4d28a198f1773a3ff21e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '012426f098ce40c3aaa00f628fe9cebb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 773.919696] env[62385]: DEBUG oslo_vmware.api [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205652, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.957599] env[62385]: DEBUG nova.compute.manager [req-d89291b5-8787-48cf-bb4d-d76736b97a30 req-6daad645-3d76-4de7-a29e-b840cb484b72 service nova] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Received event network-vif-deleted-402f6138-a737-4b40-acf1-3776bf4db1e5 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 773.957821] env[62385]: INFO nova.compute.manager [req-d89291b5-8787-48cf-bb4d-d76736b97a30 req-6daad645-3d76-4de7-a29e-b840cb484b72 service nova] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Neutron deleted interface 402f6138-a737-4b40-acf1-3776bf4db1e5; detaching it from the instance and deleting it from the info cache [ 773.958485] env[62385]: DEBUG nova.network.neutron [req-d89291b5-8787-48cf-bb4d-d76736b97a30 req-6daad645-3d76-4de7-a29e-b840cb484b72 service nova] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.142639] env[62385]: DEBUG oslo_vmware.api [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': task-1205649, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.7759} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.143117] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7/8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 774.143590] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 774.143814] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5e6facfa-670b-4fe3-a97b-a647a5f53976 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.155997] env[62385]: DEBUG oslo_vmware.api [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Waiting for the task: (returnval){ [ 774.155997] env[62385]: value = "task-1205653" [ 774.155997] env[62385]: _type = "Task" [ 774.155997] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.167613] env[62385]: DEBUG oslo_vmware.api [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Task: {'id': task-1205650, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.67352} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.171156] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] f77dc5ab-7e63-4a1b-9c87-e474a7e28c93/f77dc5ab-7e63-4a1b-9c87-e474a7e28c93.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 774.171419] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 774.172025] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2a3a483b-01e1-4d9f-ba13-2c9e85e9b336 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.177186] env[62385]: DEBUG oslo_vmware.api [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': task-1205653, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.181530] env[62385]: DEBUG oslo_vmware.api [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Waiting for the task: (returnval){ [ 774.181530] env[62385]: value = "task-1205654" [ 774.181530] env[62385]: _type = "Task" [ 774.181530] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.193474] env[62385]: DEBUG oslo_vmware.api [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Task: {'id': task-1205654, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.203993] env[62385]: DEBUG nova.network.neutron [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Updating instance_info_cache with network_info: [{"id": "b838271e-07e8-4880-980a-703c1bbd4da5", "address": "fa:16:3e:f6:2c:f9", "network": {"id": "1c117704-eb0e-43fc-84fe-5e5059cbb310", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1332686438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "395adff2d8204123b16bf9e8f5c04d05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56834f67-27a8-43dc-bbc6-a74aaa08959b", "external-id": "nsx-vlan-transportzone-949", "segmentation_id": 949, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb838271e-07", "ovs_interfaceid": "b838271e-07e8-4880-980a-703c1bbd4da5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.225082] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f1ed29c2-f0f6-4349-82f1-05918a2b3238 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Lock "e73ecefc-fed3-4ccd-88d3-399a4b72bb4b" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 3.865s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.251046] env[62385]: DEBUG nova.network.neutron [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Successfully created port: ad929652-f520-465a-aeea-1db4e8f36fc3 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 774.281008] env[62385]: DEBUG nova.network.neutron [-] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.283234] env[62385]: DEBUG oslo_concurrency.lockutils [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.349931] env[62385]: DEBUG nova.compute.manager [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 774.378178] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-581483ac-c2f7-4272-8f2a-8634783c8c86 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.386812] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98abbf5d-0742-49db-901f-2cec4bb43236 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.430107] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-267de25f-f94a-4e84-8f9c-f4caf9abc896 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.438969] env[62385]: DEBUG oslo_vmware.api [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205652, 'name': PowerOnVM_Task, 'duration_secs': 0.71673} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.441306] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 774.441552] env[62385]: INFO nova.compute.manager [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Took 8.83 seconds to spawn the instance on the hypervisor. [ 774.441748] env[62385]: DEBUG nova.compute.manager [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 774.442618] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b515a8d1-5ab0-4a18-bde8-ed0fe917caac {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.446361] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-588e1d10-8af0-4392-b822-5ad89e0d04e0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.465753] env[62385]: DEBUG nova.compute.provider_tree [None req-8b788b1c-9691-4fa5-a523-a8b477c9025c tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 774.467279] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0a8bb2ec-2036-41e8-b2cf-3960df460c59 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.479060] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3374ad88-862d-4bb2-859d-9074a0d3a0b2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.509241] env[62385]: DEBUG nova.compute.manager [req-d89291b5-8787-48cf-bb4d-d76736b97a30 req-6daad645-3d76-4de7-a29e-b840cb484b72 service nova] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Detach interface failed, port_id=402f6138-a737-4b40-acf1-3776bf4db1e5, reason: Instance 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69 could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 774.668268] env[62385]: DEBUG oslo_vmware.api [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': task-1205653, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.179754} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.668563] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 774.669439] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-068cdfe4-849b-4d14-8e6a-c13f94dbdd8b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.690371] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Reconfiguring VM instance instance-0000003a to attach disk [datastore1] 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7/8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 774.690756] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e59f0f84-7d73-4040-96ac-5882a8a8e31f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.710025] env[62385]: DEBUG oslo_concurrency.lockutils [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Releasing lock "refresh_cache-65bb1d14-84b2-4e75-acdc-dc674a035101" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.718183] env[62385]: DEBUG oslo_vmware.api [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Task: {'id': task-1205654, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.153638} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.719756] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 774.721346] env[62385]: DEBUG oslo_vmware.api [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Waiting for the task: (returnval){ [ 774.721346] env[62385]: value = "task-1205655" [ 774.721346] env[62385]: _type = "Task" [ 774.721346] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.721346] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0985177c-3af4-45dc-afb1-64aa97bce681 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.752297] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Reconfiguring VM instance instance-00000039 to attach disk [datastore2] f77dc5ab-7e63-4a1b-9c87-e474a7e28c93/f77dc5ab-7e63-4a1b-9c87-e474a7e28c93.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 774.760364] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0789028d-aa72-45d6-92d7-4fea8c9fb1d2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.777257] env[62385]: DEBUG oslo_vmware.api [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': task-1205655, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.781733] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 774.782987] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dfc8b6f7-b2a3-4723-b023-ac31a36f4447 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.788165] env[62385]: INFO nova.compute.manager [-] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Took 1.34 seconds to deallocate network for instance. [ 774.788715] env[62385]: DEBUG oslo_vmware.api [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Waiting for the task: (returnval){ [ 774.788715] env[62385]: value = "task-1205656" [ 774.788715] env[62385]: _type = "Task" [ 774.788715] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.799479] env[62385]: DEBUG oslo_vmware.api [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Waiting for the task: (returnval){ [ 774.799479] env[62385]: value = "task-1205657" [ 774.799479] env[62385]: _type = "Task" [ 774.799479] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.813557] env[62385]: DEBUG oslo_vmware.api [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Task: {'id': task-1205656, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.819099] env[62385]: DEBUG oslo_vmware.api [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': task-1205657, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.981141] env[62385]: DEBUG nova.scheduler.client.report [None req-8b788b1c-9691-4fa5-a523-a8b477c9025c tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 774.983624] env[62385]: INFO nova.compute.manager [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Took 37.38 seconds to build instance. [ 775.235451] env[62385]: DEBUG oslo_vmware.api [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': task-1205655, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.302888] env[62385]: DEBUG oslo_concurrency.lockutils [None req-181b8245-1e74-4984-9c8b-2576aab7b28f tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.302888] env[62385]: DEBUG oslo_vmware.api [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Task: {'id': task-1205656, 'name': ReconfigVM_Task, 'duration_secs': 0.410212} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.302888] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Reconfigured VM instance instance-00000039 to attach disk [datastore2] f77dc5ab-7e63-4a1b-9c87-e474a7e28c93/f77dc5ab-7e63-4a1b-9c87-e474a7e28c93.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 775.303097] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-56fa12b4-5e17-455c-9c3b-864d9ced7d91 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.309438] env[62385]: DEBUG oslo_vmware.api [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Waiting for the task: (returnval){ [ 775.309438] env[62385]: value = "task-1205658" [ 775.309438] env[62385]: _type = "Task" [ 775.309438] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.314840] env[62385]: DEBUG oslo_vmware.api [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': task-1205657, 'name': PowerOffVM_Task, 'duration_secs': 0.265209} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.316667] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 775.317437] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed7ec89c-fa23-4a27-a2eb-a20ba166834f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.325058] env[62385]: DEBUG oslo_vmware.api [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Task: {'id': task-1205658, 'name': Rename_Task} progress is 10%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.342059] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d7f1506-4272-44ab-9b2f-8fb57ef52d02 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.361579] env[62385]: DEBUG nova.compute.manager [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 775.381884] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 775.383602] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5174c5a0-9c2c-4939-aa4a-366752b856fe {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.391328] env[62385]: DEBUG nova.virt.hardware [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 775.391591] env[62385]: DEBUG nova.virt.hardware [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 775.391756] env[62385]: DEBUG nova.virt.hardware [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 775.391942] env[62385]: DEBUG nova.virt.hardware [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 775.392113] env[62385]: DEBUG nova.virt.hardware [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 775.392278] env[62385]: DEBUG nova.virt.hardware [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 775.392533] env[62385]: DEBUG nova.virt.hardware [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 775.392718] env[62385]: DEBUG nova.virt.hardware [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 775.392895] env[62385]: DEBUG nova.virt.hardware [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 775.393077] env[62385]: DEBUG nova.virt.hardware [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 775.393261] env[62385]: DEBUG nova.virt.hardware [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 775.394138] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d47d1bf-d3bb-4743-9a89-ee38f16a79d8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.399507] env[62385]: DEBUG oslo_vmware.api [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Waiting for the task: (returnval){ [ 775.399507] env[62385]: value = "task-1205659" [ 775.399507] env[62385]: _type = "Task" [ 775.399507] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.401038] env[62385]: DEBUG nova.objects.instance [None req-17f52ec1-e766-418f-9f9c-efabfaf02d12 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Lazy-loading 'flavor' on Instance uuid f50ddafe-f5a9-422b-b0f4-46c1b111dfbe {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 775.410355] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa5cfcea-d2e1-4de8-85c0-6a62dde7893d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.421140] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] VM already powered off {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 775.421140] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 775.421140] env[62385]: DEBUG oslo_concurrency.lockutils [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 775.421140] env[62385]: DEBUG oslo_concurrency.lockutils [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.421550] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 775.428629] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-561824fa-16b5-4116-b5c5-a6ec144e2eb1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.439039] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 775.439039] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 775.439718] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8605a3f7-2b09-4e66-bb68-073cf2e04230 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.445187] env[62385]: DEBUG oslo_vmware.api [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Waiting for the task: (returnval){ [ 775.445187] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52747cb6-8c8a-5099-71ff-0d9502efd6fb" [ 775.445187] env[62385]: _type = "Task" [ 775.445187] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.457837] env[62385]: DEBUG oslo_vmware.api [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52747cb6-8c8a-5099-71ff-0d9502efd6fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.486607] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8b788b1c-9691-4fa5-a523-a8b477c9025c tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.147s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.488130] env[62385]: DEBUG oslo_concurrency.lockutils [None req-12275953-4460-4b25-b23b-439197a51a3f tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "f8a98710-9d02-4aa3-b6a5-34b7060ad62d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.086s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.488615] env[62385]: DEBUG oslo_concurrency.lockutils [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.755s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.490158] env[62385]: INFO nova.compute.claims [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 775.511567] env[62385]: INFO nova.scheduler.client.report [None req-8b788b1c-9691-4fa5-a523-a8b477c9025c tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Deleted allocations for instance 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee [ 775.710832] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "f8a98710-9d02-4aa3-b6a5-34b7060ad62d" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.710981] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "f8a98710-9d02-4aa3-b6a5-34b7060ad62d" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.711152] env[62385]: INFO nova.compute.manager [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Shelving [ 775.735588] env[62385]: DEBUG oslo_vmware.api [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': task-1205655, 'name': ReconfigVM_Task, 'duration_secs': 0.545078} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.735588] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Reconfigured VM instance instance-0000003a to attach disk [datastore1] 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7/8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 775.736071] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-006a42cb-e26f-43d8-a9ef-9171e64ef5bd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.744235] env[62385]: DEBUG oslo_vmware.api [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Waiting for the task: (returnval){ [ 775.744235] env[62385]: value = "task-1205660" [ 775.744235] env[62385]: _type = "Task" [ 775.744235] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.753465] env[62385]: DEBUG oslo_vmware.api [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': task-1205660, 'name': Rename_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.825628] env[62385]: DEBUG oslo_vmware.api [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Task: {'id': task-1205658, 'name': Rename_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.909213] env[62385]: DEBUG oslo_concurrency.lockutils [None req-17f52ec1-e766-418f-9f9c-efabfaf02d12 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Acquiring lock "refresh_cache-f50ddafe-f5a9-422b-b0f4-46c1b111dfbe" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 775.909502] env[62385]: DEBUG oslo_concurrency.lockutils [None req-17f52ec1-e766-418f-9f9c-efabfaf02d12 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Acquired lock "refresh_cache-f50ddafe-f5a9-422b-b0f4-46c1b111dfbe" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.956737] env[62385]: DEBUG oslo_vmware.api [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52747cb6-8c8a-5099-71ff-0d9502efd6fb, 'name': SearchDatastore_Task, 'duration_secs': 0.011667} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.957595] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-298d1322-6d51-40d8-b40a-179550d9e4e1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.964557] env[62385]: DEBUG oslo_vmware.api [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Waiting for the task: (returnval){ [ 775.964557] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]520aaaab-72ca-50c4-f67c-050f593f226e" [ 775.964557] env[62385]: _type = "Task" [ 775.964557] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.973800] env[62385]: DEBUG oslo_vmware.api [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]520aaaab-72ca-50c4-f67c-050f593f226e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.999097] env[62385]: DEBUG nova.compute.manager [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 776.008724] env[62385]: DEBUG nova.compute.manager [req-0188b7c2-d374-4ce8-be28-e356e0e046a2 req-08ec7595-f547-40f9-b290-fc98ed6de689 service nova] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Received event network-vif-plugged-ad929652-f520-465a-aeea-1db4e8f36fc3 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 776.009033] env[62385]: DEBUG oslo_concurrency.lockutils [req-0188b7c2-d374-4ce8-be28-e356e0e046a2 req-08ec7595-f547-40f9-b290-fc98ed6de689 service nova] Acquiring lock "0d823fcc-cbba-41f6-a47c-2f57d3e79948-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.009326] env[62385]: DEBUG oslo_concurrency.lockutils [req-0188b7c2-d374-4ce8-be28-e356e0e046a2 req-08ec7595-f547-40f9-b290-fc98ed6de689 service nova] Lock "0d823fcc-cbba-41f6-a47c-2f57d3e79948-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.009543] env[62385]: DEBUG oslo_concurrency.lockutils [req-0188b7c2-d374-4ce8-be28-e356e0e046a2 req-08ec7595-f547-40f9-b290-fc98ed6de689 service nova] Lock "0d823fcc-cbba-41f6-a47c-2f57d3e79948-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.009736] env[62385]: DEBUG nova.compute.manager [req-0188b7c2-d374-4ce8-be28-e356e0e046a2 req-08ec7595-f547-40f9-b290-fc98ed6de689 service nova] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] No waiting events found dispatching network-vif-plugged-ad929652-f520-465a-aeea-1db4e8f36fc3 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 776.010037] env[62385]: WARNING nova.compute.manager [req-0188b7c2-d374-4ce8-be28-e356e0e046a2 req-08ec7595-f547-40f9-b290-fc98ed6de689 service nova] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Received unexpected event network-vif-plugged-ad929652-f520-465a-aeea-1db4e8f36fc3 for instance with vm_state building and task_state spawning. [ 776.020408] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8b788b1c-9691-4fa5-a523-a8b477c9025c tempest-ServerPasswordTestJSON-567585234 tempest-ServerPasswordTestJSON-567585234-project-member] Lock "8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.588s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.221699] env[62385]: DEBUG nova.compute.manager [req-85268404-38f3-417f-add5-c9ec9d201fb0 req-eac452b8-a137-41fd-bb47-4ad460f49696 service nova] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Received event network-changed-30e19299-99d0-4c98-8615-78f2f5f76dd5 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 776.221983] env[62385]: DEBUG nova.compute.manager [req-85268404-38f3-417f-add5-c9ec9d201fb0 req-eac452b8-a137-41fd-bb47-4ad460f49696 service nova] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Refreshing instance network info cache due to event network-changed-30e19299-99d0-4c98-8615-78f2f5f76dd5. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 776.222320] env[62385]: DEBUG oslo_concurrency.lockutils [req-85268404-38f3-417f-add5-c9ec9d201fb0 req-eac452b8-a137-41fd-bb47-4ad460f49696 service nova] Acquiring lock "refresh_cache-e73ecefc-fed3-4ccd-88d3-399a4b72bb4b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.222545] env[62385]: DEBUG oslo_concurrency.lockutils [req-85268404-38f3-417f-add5-c9ec9d201fb0 req-eac452b8-a137-41fd-bb47-4ad460f49696 service nova] Acquired lock "refresh_cache-e73ecefc-fed3-4ccd-88d3-399a4b72bb4b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.222805] env[62385]: DEBUG nova.network.neutron [req-85268404-38f3-417f-add5-c9ec9d201fb0 req-eac452b8-a137-41fd-bb47-4ad460f49696 service nova] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Refreshing network info cache for port 30e19299-99d0-4c98-8615-78f2f5f76dd5 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 776.230776] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 776.231265] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8a781442-2c9f-4f5e-a9cb-dcdee852ea9d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.240634] env[62385]: DEBUG oslo_vmware.api [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 776.240634] env[62385]: value = "task-1205661" [ 776.240634] env[62385]: _type = "Task" [ 776.240634] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.254211] env[62385]: DEBUG oslo_vmware.api [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205661, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.257444] env[62385]: DEBUG oslo_vmware.api [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': task-1205660, 'name': Rename_Task, 'duration_secs': 0.220241} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.257792] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 776.258123] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6030b5ba-fe4d-44c6-97a0-9dedd8bbe6e9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.266579] env[62385]: DEBUG oslo_vmware.api [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Waiting for the task: (returnval){ [ 776.266579] env[62385]: value = "task-1205662" [ 776.266579] env[62385]: _type = "Task" [ 776.266579] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.276982] env[62385]: DEBUG oslo_vmware.api [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': task-1205662, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.312294] env[62385]: DEBUG nova.network.neutron [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Successfully updated port: ad929652-f520-465a-aeea-1db4e8f36fc3 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 776.329189] env[62385]: DEBUG oslo_vmware.api [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Task: {'id': task-1205658, 'name': Rename_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.477234] env[62385]: DEBUG oslo_vmware.api [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]520aaaab-72ca-50c4-f67c-050f593f226e, 'name': SearchDatastore_Task, 'duration_secs': 0.012251} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.477627] env[62385]: DEBUG oslo_concurrency.lockutils [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.477933] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 65bb1d14-84b2-4e75-acdc-dc674a035101/c0abbb8d-77e7-4f22-a256-0faf0a781109-rescue.vmdk. {{(pid=62385) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 776.478360] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b0ca4b97-7dce-48d7-8a58-2e86c32b5c62 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.491457] env[62385]: DEBUG oslo_vmware.api [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Waiting for the task: (returnval){ [ 776.491457] env[62385]: value = "task-1205663" [ 776.491457] env[62385]: _type = "Task" [ 776.491457] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.509445] env[62385]: DEBUG oslo_vmware.api [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': task-1205663, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.515707] env[62385]: DEBUG nova.compute.manager [req-1843a952-f365-49e0-8858-405fae6bf5d6 req-d32d0db1-e47e-49ea-9d68-0d76723969d2 service nova] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Received event network-changed-ad929652-f520-465a-aeea-1db4e8f36fc3 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 776.515945] env[62385]: DEBUG nova.compute.manager [req-1843a952-f365-49e0-8858-405fae6bf5d6 req-d32d0db1-e47e-49ea-9d68-0d76723969d2 service nova] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Refreshing instance network info cache due to event network-changed-ad929652-f520-465a-aeea-1db4e8f36fc3. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 776.516244] env[62385]: DEBUG oslo_concurrency.lockutils [req-1843a952-f365-49e0-8858-405fae6bf5d6 req-d32d0db1-e47e-49ea-9d68-0d76723969d2 service nova] Acquiring lock "refresh_cache-0d823fcc-cbba-41f6-a47c-2f57d3e79948" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.516447] env[62385]: DEBUG oslo_concurrency.lockutils [req-1843a952-f365-49e0-8858-405fae6bf5d6 req-d32d0db1-e47e-49ea-9d68-0d76723969d2 service nova] Acquired lock "refresh_cache-0d823fcc-cbba-41f6-a47c-2f57d3e79948" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.516645] env[62385]: DEBUG nova.network.neutron [req-1843a952-f365-49e0-8858-405fae6bf5d6 req-d32d0db1-e47e-49ea-9d68-0d76723969d2 service nova] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Refreshing network info cache for port ad929652-f520-465a-aeea-1db4e8f36fc3 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 776.536989] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.604745] env[62385]: DEBUG nova.network.neutron [None req-17f52ec1-e766-418f-9f9c-efabfaf02d12 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 776.755149] env[62385]: DEBUG oslo_vmware.api [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205661, 'name': PowerOffVM_Task, 'duration_secs': 0.31644} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.755450] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 776.756302] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7169729f-b3ec-41e4-9827-869424e02318 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.786892] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d47dd32-f09f-4b77-b322-09babe08148b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.798981] env[62385]: DEBUG oslo_vmware.api [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': task-1205662, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.821438] env[62385]: DEBUG oslo_concurrency.lockutils [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "refresh_cache-0d823fcc-cbba-41f6-a47c-2f57d3e79948" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.829537] env[62385]: DEBUG oslo_vmware.api [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Task: {'id': task-1205658, 'name': Rename_Task, 'duration_secs': 1.286539} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.832784] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 776.833324] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5cf79fda-2d03-479f-adbb-db0211e5290e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.845640] env[62385]: DEBUG oslo_vmware.api [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Waiting for the task: (returnval){ [ 776.845640] env[62385]: value = "task-1205664" [ 776.845640] env[62385]: _type = "Task" [ 776.845640] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.856741] env[62385]: DEBUG oslo_vmware.api [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Task: {'id': task-1205664, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.999793] env[62385]: DEBUG oslo_vmware.api [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': task-1205663, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.056939] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0603f934-77b7-4ffd-b324-712bf3a3f225 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.067318] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d027157b-0981-4456-8d46-05769af9b108 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.072141] env[62385]: DEBUG nova.network.neutron [req-1843a952-f365-49e0-8858-405fae6bf5d6 req-d32d0db1-e47e-49ea-9d68-0d76723969d2 service nova] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 777.115383] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e12f07a7-53f7-479d-839e-8ea66f49580d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.129410] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84bee1c7-232b-4df1-aef5-46c3814253ef {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.146871] env[62385]: DEBUG nova.compute.provider_tree [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 777.194305] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b6753bc4-02bf-4300-903f-acd56a4146b9 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Acquiring lock "e73ecefc-fed3-4ccd-88d3-399a4b72bb4b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.194745] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b6753bc4-02bf-4300-903f-acd56a4146b9 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Lock "e73ecefc-fed3-4ccd-88d3-399a4b72bb4b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.195013] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b6753bc4-02bf-4300-903f-acd56a4146b9 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Acquiring lock "e73ecefc-fed3-4ccd-88d3-399a4b72bb4b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.195304] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b6753bc4-02bf-4300-903f-acd56a4146b9 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Lock "e73ecefc-fed3-4ccd-88d3-399a4b72bb4b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.195534] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b6753bc4-02bf-4300-903f-acd56a4146b9 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Lock "e73ecefc-fed3-4ccd-88d3-399a4b72bb4b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.198012] env[62385]: INFO nova.compute.manager [None req-b6753bc4-02bf-4300-903f-acd56a4146b9 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Terminating instance [ 777.201021] env[62385]: DEBUG nova.compute.manager [None req-b6753bc4-02bf-4300-903f-acd56a4146b9 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 777.201230] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-b6753bc4-02bf-4300-903f-acd56a4146b9 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 777.202412] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61b0ce87-79a1-4a39-9a99-82689f7d1a1f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.213694] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6753bc4-02bf-4300-903f-acd56a4146b9 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 777.213986] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ce983c0e-fdec-43df-8d3e-9253cad6fadb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.217226] env[62385]: DEBUG nova.network.neutron [req-1843a952-f365-49e0-8858-405fae6bf5d6 req-d32d0db1-e47e-49ea-9d68-0d76723969d2 service nova] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.224773] env[62385]: DEBUG oslo_vmware.api [None req-b6753bc4-02bf-4300-903f-acd56a4146b9 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Waiting for the task: (returnval){ [ 777.224773] env[62385]: value = "task-1205665" [ 777.224773] env[62385]: _type = "Task" [ 777.224773] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.240584] env[62385]: DEBUG oslo_vmware.api [None req-b6753bc4-02bf-4300-903f-acd56a4146b9 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Task: {'id': task-1205665, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.249646] env[62385]: DEBUG nova.network.neutron [req-85268404-38f3-417f-add5-c9ec9d201fb0 req-eac452b8-a137-41fd-bb47-4ad460f49696 service nova] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Updated VIF entry in instance network info cache for port 30e19299-99d0-4c98-8615-78f2f5f76dd5. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 777.250032] env[62385]: DEBUG nova.network.neutron [req-85268404-38f3-417f-add5-c9ec9d201fb0 req-eac452b8-a137-41fd-bb47-4ad460f49696 service nova] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Updating instance_info_cache with network_info: [{"id": "30e19299-99d0-4c98-8615-78f2f5f76dd5", "address": "fa:16:3e:a9:58:3d", "network": {"id": "63367e0f-9855-478b-a546-5f0cf8190519", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-21852824-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ebc1c1ed0bb404b89e59325535edecd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "778b9a40-d603-4765-ac88-bd6d42c457a2", "external-id": "nsx-vlan-transportzone-114", "segmentation_id": 114, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30e19299-99", "ovs_interfaceid": "30e19299-99d0-4c98-8615-78f2f5f76dd5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.291824] env[62385]: DEBUG oslo_vmware.api [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': task-1205662, 'name': PowerOnVM_Task, 'duration_secs': 0.888572} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.291824] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 777.291975] env[62385]: INFO nova.compute.manager [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Took 6.46 seconds to spawn the instance on the hypervisor. [ 777.292203] env[62385]: DEBUG nova.compute.manager [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 777.292968] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c15fe11-fcbf-496f-85d7-586e4f3987c9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.311287] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Creating Snapshot of the VM instance {{(pid=62385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 777.311287] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-d9f829e1-0f6a-42a9-91d5-436f464229d5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.320062] env[62385]: DEBUG oslo_vmware.api [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 777.320062] env[62385]: value = "task-1205666" [ 777.320062] env[62385]: _type = "Task" [ 777.320062] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.336335] env[62385]: DEBUG oslo_vmware.api [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205666, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.359959] env[62385]: DEBUG oslo_vmware.api [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Task: {'id': task-1205664, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.503269] env[62385]: DEBUG oslo_vmware.api [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': task-1205663, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.616857} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.503269] env[62385]: INFO nova.virt.vmwareapi.ds_util [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 65bb1d14-84b2-4e75-acdc-dc674a035101/c0abbb8d-77e7-4f22-a256-0faf0a781109-rescue.vmdk. [ 777.504329] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08367c55-63b4-45d5-b0e2-cbef5bfa228e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.533763] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Reconfiguring VM instance instance-00000037 to attach disk [datastore2] 65bb1d14-84b2-4e75-acdc-dc674a035101/c0abbb8d-77e7-4f22-a256-0faf0a781109-rescue.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 777.537809] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-40b2ca41-257b-4244-97ff-489a0620f215 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.562844] env[62385]: DEBUG oslo_vmware.api [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Waiting for the task: (returnval){ [ 777.562844] env[62385]: value = "task-1205667" [ 777.562844] env[62385]: _type = "Task" [ 777.562844] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.569588] env[62385]: DEBUG oslo_vmware.api [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': task-1205667, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.652698] env[62385]: DEBUG nova.scheduler.client.report [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 777.673138] env[62385]: DEBUG nova.network.neutron [None req-17f52ec1-e766-418f-9f9c-efabfaf02d12 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Updating instance_info_cache with network_info: [{"id": "fdae3362-d2d0-4a5e-80ff-c92e442ef9f4", "address": "fa:16:3e:87:99:cf", "network": {"id": "6a072c47-f3cd-4e88-a6f4-bb03318b678f", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1088193534-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}, {"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.199", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ff114d055f440e0afc6c87e2ab6b4e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b5c60ce-845e-4506-bc10-348461fece6d", "external-id": "nsx-vlan-transportzone-831", "segmentation_id": 831, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfdae3362-d2", "ovs_interfaceid": "fdae3362-d2d0-4a5e-80ff-c92e442ef9f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.723408] env[62385]: DEBUG oslo_concurrency.lockutils [req-1843a952-f365-49e0-8858-405fae6bf5d6 req-d32d0db1-e47e-49ea-9d68-0d76723969d2 service nova] Releasing lock "refresh_cache-0d823fcc-cbba-41f6-a47c-2f57d3e79948" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.723920] env[62385]: DEBUG oslo_concurrency.lockutils [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquired lock "refresh_cache-0d823fcc-cbba-41f6-a47c-2f57d3e79948" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.725030] env[62385]: DEBUG nova.network.neutron [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 777.737281] env[62385]: DEBUG oslo_vmware.api [None req-b6753bc4-02bf-4300-903f-acd56a4146b9 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Task: {'id': task-1205665, 'name': PowerOffVM_Task, 'duration_secs': 0.241811} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.737668] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6753bc4-02bf-4300-903f-acd56a4146b9 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 777.737853] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-b6753bc4-02bf-4300-903f-acd56a4146b9 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 777.738144] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7c1437fe-220d-457c-ab4a-5d72d0e8b4ab {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.753716] env[62385]: DEBUG oslo_concurrency.lockutils [req-85268404-38f3-417f-add5-c9ec9d201fb0 req-eac452b8-a137-41fd-bb47-4ad460f49696 service nova] Releasing lock "refresh_cache-e73ecefc-fed3-4ccd-88d3-399a4b72bb4b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.819356] env[62385]: INFO nova.compute.manager [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Took 34.93 seconds to build instance. [ 777.834053] env[62385]: DEBUG oslo_vmware.api [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205666, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.857660] env[62385]: DEBUG oslo_vmware.api [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Task: {'id': task-1205664, 'name': PowerOnVM_Task, 'duration_secs': 0.696965} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.857938] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 777.858244] env[62385]: INFO nova.compute.manager [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Took 9.69 seconds to spawn the instance on the hypervisor. [ 777.858453] env[62385]: DEBUG nova.compute.manager [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 777.859325] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfbdf66d-3ee2-46e6-bedd-b0b2f775d96f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.988770] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-b6753bc4-02bf-4300-903f-acd56a4146b9 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 777.989126] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-b6753bc4-02bf-4300-903f-acd56a4146b9 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 777.989431] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-b6753bc4-02bf-4300-903f-acd56a4146b9 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Deleting the datastore file [datastore1] e73ecefc-fed3-4ccd-88d3-399a4b72bb4b {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 777.990122] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-137e2721-88ec-4e2c-9d86-f07b1b8ff743 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.008369] env[62385]: DEBUG oslo_vmware.api [None req-b6753bc4-02bf-4300-903f-acd56a4146b9 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Waiting for the task: (returnval){ [ 778.008369] env[62385]: value = "task-1205669" [ 778.008369] env[62385]: _type = "Task" [ 778.008369] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.018686] env[62385]: DEBUG oslo_vmware.api [None req-b6753bc4-02bf-4300-903f-acd56a4146b9 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Task: {'id': task-1205669, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.071101] env[62385]: DEBUG oslo_vmware.api [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': task-1205667, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.160383] env[62385]: DEBUG oslo_concurrency.lockutils [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.672s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.161079] env[62385]: DEBUG nova.compute.manager [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 778.164379] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.732s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 778.165320] env[62385]: INFO nova.compute.claims [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 778.178284] env[62385]: DEBUG oslo_concurrency.lockutils [None req-17f52ec1-e766-418f-9f9c-efabfaf02d12 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Releasing lock "refresh_cache-f50ddafe-f5a9-422b-b0f4-46c1b111dfbe" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.178557] env[62385]: DEBUG nova.compute.manager [None req-17f52ec1-e766-418f-9f9c-efabfaf02d12 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Inject network info {{(pid=62385) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7219}} [ 778.178841] env[62385]: DEBUG nova.compute.manager [None req-17f52ec1-e766-418f-9f9c-efabfaf02d12 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] network_info to inject: |[{"id": "fdae3362-d2d0-4a5e-80ff-c92e442ef9f4", "address": "fa:16:3e:87:99:cf", "network": {"id": "6a072c47-f3cd-4e88-a6f4-bb03318b678f", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1088193534-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}, {"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.199", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ff114d055f440e0afc6c87e2ab6b4e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b5c60ce-845e-4506-bc10-348461fece6d", "external-id": "nsx-vlan-transportzone-831", "segmentation_id": 831, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfdae3362-d2", "ovs_interfaceid": "fdae3362-d2d0-4a5e-80ff-c92e442ef9f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7220}} [ 778.185941] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-17f52ec1-e766-418f-9f9c-efabfaf02d12 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Reconfiguring VM instance to set the machine id {{(pid=62385) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1802}} [ 778.185941] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c4b31a04-e66a-4a30-8d3a-7f3b275cbf9f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.202377] env[62385]: DEBUG oslo_vmware.api [None req-17f52ec1-e766-418f-9f9c-efabfaf02d12 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Waiting for the task: (returnval){ [ 778.202377] env[62385]: value = "task-1205670" [ 778.202377] env[62385]: _type = "Task" [ 778.202377] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.212541] env[62385]: DEBUG oslo_vmware.api [None req-17f52ec1-e766-418f-9f9c-efabfaf02d12 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Task: {'id': task-1205670, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.296440] env[62385]: DEBUG nova.network.neutron [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 778.324696] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bf357749-82e1-4572-8151-e5b07fbec9cb tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Lock "8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 105.383s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.332327] env[62385]: DEBUG oslo_vmware.api [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205666, 'name': CreateSnapshot_Task, 'duration_secs': 0.93266} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.332327] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Created Snapshot of the VM instance {{(pid=62385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 778.332820] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc966dc1-7489-42e0-85f9-2cdb5a9facc8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.377900] env[62385]: INFO nova.compute.manager [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Took 35.67 seconds to build instance. [ 778.392920] env[62385]: DEBUG nova.objects.instance [None req-eb8f459c-5746-44f6-b3e4-e73b2e5a968e tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Lazy-loading 'flavor' on Instance uuid f50ddafe-f5a9-422b-b0f4-46c1b111dfbe {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 778.417212] env[62385]: DEBUG nova.compute.manager [req-20c6cdad-446d-4e8c-a9f6-505efc5e40e5 req-7ad23db0-8e83-4a28-a035-6dccba0019b8 service nova] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Received event network-changed-fdae3362-d2d0-4a5e-80ff-c92e442ef9f4 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 778.417212] env[62385]: DEBUG nova.compute.manager [req-20c6cdad-446d-4e8c-a9f6-505efc5e40e5 req-7ad23db0-8e83-4a28-a035-6dccba0019b8 service nova] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Refreshing instance network info cache due to event network-changed-fdae3362-d2d0-4a5e-80ff-c92e442ef9f4. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 778.417212] env[62385]: DEBUG oslo_concurrency.lockutils [req-20c6cdad-446d-4e8c-a9f6-505efc5e40e5 req-7ad23db0-8e83-4a28-a035-6dccba0019b8 service nova] Acquiring lock "refresh_cache-f50ddafe-f5a9-422b-b0f4-46c1b111dfbe" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.417212] env[62385]: DEBUG oslo_concurrency.lockutils [req-20c6cdad-446d-4e8c-a9f6-505efc5e40e5 req-7ad23db0-8e83-4a28-a035-6dccba0019b8 service nova] Acquired lock "refresh_cache-f50ddafe-f5a9-422b-b0f4-46c1b111dfbe" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.417212] env[62385]: DEBUG nova.network.neutron [req-20c6cdad-446d-4e8c-a9f6-505efc5e40e5 req-7ad23db0-8e83-4a28-a035-6dccba0019b8 service nova] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Refreshing network info cache for port fdae3362-d2d0-4a5e-80ff-c92e442ef9f4 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 778.491124] env[62385]: INFO nova.compute.manager [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Rebuilding instance [ 778.521649] env[62385]: DEBUG oslo_vmware.api [None req-b6753bc4-02bf-4300-903f-acd56a4146b9 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Task: {'id': task-1205669, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.317905} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.527939] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-b6753bc4-02bf-4300-903f-acd56a4146b9 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 778.527939] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-b6753bc4-02bf-4300-903f-acd56a4146b9 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 778.528244] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-b6753bc4-02bf-4300-903f-acd56a4146b9 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 778.528244] env[62385]: INFO nova.compute.manager [None req-b6753bc4-02bf-4300-903f-acd56a4146b9 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Took 1.33 seconds to destroy the instance on the hypervisor. [ 778.528510] env[62385]: DEBUG oslo.service.loopingcall [None req-b6753bc4-02bf-4300-903f-acd56a4146b9 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 778.528920] env[62385]: DEBUG nova.compute.manager [-] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 778.529022] env[62385]: DEBUG nova.network.neutron [-] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 778.550705] env[62385]: DEBUG nova.compute.manager [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 778.551573] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5cb0801-f4bf-470f-a5c1-075c60650719 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.569923] env[62385]: DEBUG oslo_vmware.api [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': task-1205667, 'name': ReconfigVM_Task, 'duration_secs': 0.663368} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.570210] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Reconfigured VM instance instance-00000037 to attach disk [datastore2] 65bb1d14-84b2-4e75-acdc-dc674a035101/c0abbb8d-77e7-4f22-a256-0faf0a781109-rescue.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 778.571145] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a2c301b-3094-4c45-a7d2-fa131bc6b7a9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.597884] env[62385]: DEBUG nova.network.neutron [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Updating instance_info_cache with network_info: [{"id": "ad929652-f520-465a-aeea-1db4e8f36fc3", "address": "fa:16:3e:b6:00:7c", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad929652-f5", "ovs_interfaceid": "ad929652-f520-465a-aeea-1db4e8f36fc3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.599531] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fc3e5f15-6a2e-4ba6-90d8-b360462bc5d4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.622914] env[62385]: DEBUG oslo_vmware.api [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Waiting for the task: (returnval){ [ 778.622914] env[62385]: value = "task-1205671" [ 778.622914] env[62385]: _type = "Task" [ 778.622914] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.631521] env[62385]: DEBUG oslo_vmware.api [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': task-1205671, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.673206] env[62385]: DEBUG nova.compute.utils [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 778.676617] env[62385]: DEBUG nova.compute.manager [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 778.676789] env[62385]: DEBUG nova.network.neutron [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 778.713627] env[62385]: DEBUG oslo_vmware.api [None req-17f52ec1-e766-418f-9f9c-efabfaf02d12 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Task: {'id': task-1205670, 'name': ReconfigVM_Task, 'duration_secs': 0.215334} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.714050] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-17f52ec1-e766-418f-9f9c-efabfaf02d12 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Reconfigured VM instance to set the machine id {{(pid=62385) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1805}} [ 778.827731] env[62385]: DEBUG nova.compute.manager [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 778.853476] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Creating linked-clone VM from snapshot {{(pid=62385) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 778.853882] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-2abfd569-f4ab-41ff-9c21-b2d474aca7ac {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.863789] env[62385]: DEBUG oslo_vmware.api [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 778.863789] env[62385]: value = "task-1205672" [ 778.863789] env[62385]: _type = "Task" [ 778.863789] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.876367] env[62385]: DEBUG oslo_vmware.api [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205672, 'name': CloneVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.880238] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ada07c3-d9da-41b2-9cd3-65c2c63e098f tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Lock "f77dc5ab-7e63-4a1b-9c87-e474a7e28c93" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 106.047s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.901789] env[62385]: DEBUG oslo_concurrency.lockutils [None req-eb8f459c-5746-44f6-b3e4-e73b2e5a968e tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Acquiring lock "refresh_cache-f50ddafe-f5a9-422b-b0f4-46c1b111dfbe" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.014248] env[62385]: DEBUG nova.policy [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e12425b6f5f84ca7a8ace6cb0fdadb0c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8eaa81326f084441a87ac05796958abb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 779.066339] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 779.069217] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-11bd87be-cbd9-4116-a1f3-3c3786f5f7b7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.077541] env[62385]: DEBUG oslo_vmware.api [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Waiting for the task: (returnval){ [ 779.077541] env[62385]: value = "task-1205673" [ 779.077541] env[62385]: _type = "Task" [ 779.077541] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.094492] env[62385]: DEBUG oslo_vmware.api [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': task-1205673, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.117111] env[62385]: DEBUG oslo_concurrency.lockutils [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Releasing lock "refresh_cache-0d823fcc-cbba-41f6-a47c-2f57d3e79948" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.117111] env[62385]: DEBUG nova.compute.manager [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Instance network_info: |[{"id": "ad929652-f520-465a-aeea-1db4e8f36fc3", "address": "fa:16:3e:b6:00:7c", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad929652-f5", "ovs_interfaceid": "ad929652-f520-465a-aeea-1db4e8f36fc3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 779.117294] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b6:00:7c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '685b4083-b748-41fb-a68a-273b1073fa28', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ad929652-f520-465a-aeea-1db4e8f36fc3', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 779.126448] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Creating folder: Project (012426f098ce40c3aaa00f628fe9cebb). Parent ref: group-v261107. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 779.127308] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-76041675-c9d7-42b7-b5d7-960995f513ea {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.143421] env[62385]: DEBUG oslo_vmware.api [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': task-1205671, 'name': ReconfigVM_Task, 'duration_secs': 0.347494} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.148533] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 779.150656] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-406f3123-674e-4a5e-a2b0-9abffe3b0ec0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.153183] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Created folder: Project (012426f098ce40c3aaa00f628fe9cebb) in parent group-v261107. [ 779.153404] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Creating folder: Instances. Parent ref: group-v261173. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 779.153659] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b6c4ce58-3c8d-4dd9-b6bd-8c3c08cfd583 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.162586] env[62385]: DEBUG oslo_vmware.api [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Waiting for the task: (returnval){ [ 779.162586] env[62385]: value = "task-1205675" [ 779.162586] env[62385]: _type = "Task" [ 779.162586] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.168947] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Created folder: Instances in parent group-v261173. [ 779.169284] env[62385]: DEBUG oslo.service.loopingcall [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 779.169516] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 779.169744] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-aac74deb-22fe-465f-8242-b4183f2c05a0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.188819] env[62385]: DEBUG nova.compute.manager [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 779.197261] env[62385]: DEBUG oslo_vmware.api [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': task-1205675, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.207259] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 779.207259] env[62385]: value = "task-1205677" [ 779.207259] env[62385]: _type = "Task" [ 779.207259] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.224488] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205677, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.267932] env[62385]: DEBUG nova.network.neutron [req-20c6cdad-446d-4e8c-a9f6-505efc5e40e5 req-7ad23db0-8e83-4a28-a035-6dccba0019b8 service nova] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Updated VIF entry in instance network info cache for port fdae3362-d2d0-4a5e-80ff-c92e442ef9f4. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 779.268503] env[62385]: DEBUG nova.network.neutron [req-20c6cdad-446d-4e8c-a9f6-505efc5e40e5 req-7ad23db0-8e83-4a28-a035-6dccba0019b8 service nova] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Updating instance_info_cache with network_info: [{"id": "fdae3362-d2d0-4a5e-80ff-c92e442ef9f4", "address": "fa:16:3e:87:99:cf", "network": {"id": "6a072c47-f3cd-4e88-a6f4-bb03318b678f", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1088193534-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}, {"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.199", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ff114d055f440e0afc6c87e2ab6b4e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b5c60ce-845e-4506-bc10-348461fece6d", "external-id": "nsx-vlan-transportzone-831", "segmentation_id": 831, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfdae3362-d2", "ovs_interfaceid": "fdae3362-d2d0-4a5e-80ff-c92e442ef9f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.353931] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.385620] env[62385]: DEBUG nova.compute.manager [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 779.389223] env[62385]: DEBUG oslo_vmware.api [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205672, 'name': CloneVM_Task} progress is 94%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.590583] env[62385]: DEBUG oslo_vmware.api [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': task-1205673, 'name': PowerOffVM_Task, 'duration_secs': 0.331643} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.590924] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 779.591225] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 779.592132] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9baac6b-bc5d-40cc-84f8-afe86d1a0ab3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.600540] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 779.600815] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1b677fb7-f3ee-45d6-8569-e067a8c5612c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.609996] env[62385]: DEBUG nova.network.neutron [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Successfully created port: 854fc77a-662c-4fbe-a30c-5726d1ed6088 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 779.634383] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 779.634612] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 779.634801] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Deleting the datastore file [datastore1] 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 779.635184] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a1a387df-1dae-4134-931d-5ddf8b9917fe {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.644273] env[62385]: DEBUG oslo_vmware.api [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Waiting for the task: (returnval){ [ 779.644273] env[62385]: value = "task-1205679" [ 779.644273] env[62385]: _type = "Task" [ 779.644273] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.653743] env[62385]: DEBUG oslo_vmware.api [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': task-1205679, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.676443] env[62385]: DEBUG oslo_vmware.api [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': task-1205675, 'name': PowerOnVM_Task} progress is 90%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.685324] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf879eed-20c8-43e2-970d-00a1f27a00f1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.693658] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7dbab95-3522-4176-ae8f-83e8f967ecf6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.730563] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e20581d-0a3a-4781-a135-e36dd8fb7037 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.743286] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8f31abf-cb4a-4d1f-ace9-f99156837ec0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.747249] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205677, 'name': CreateVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.760030] env[62385]: DEBUG nova.compute.provider_tree [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 779.771281] env[62385]: DEBUG oslo_concurrency.lockutils [req-20c6cdad-446d-4e8c-a9f6-505efc5e40e5 req-7ad23db0-8e83-4a28-a035-6dccba0019b8 service nova] Releasing lock "refresh_cache-f50ddafe-f5a9-422b-b0f4-46c1b111dfbe" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.771841] env[62385]: DEBUG oslo_concurrency.lockutils [None req-eb8f459c-5746-44f6-b3e4-e73b2e5a968e tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Acquired lock "refresh_cache-f50ddafe-f5a9-422b-b0f4-46c1b111dfbe" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.856230] env[62385]: DEBUG nova.network.neutron [-] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.881730] env[62385]: DEBUG oslo_vmware.api [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205672, 'name': CloneVM_Task} progress is 94%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.914744] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.154484] env[62385]: DEBUG oslo_vmware.api [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': task-1205679, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154886} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.154868] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 780.155160] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 780.155419] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 780.175911] env[62385]: DEBUG oslo_vmware.api [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': task-1205675, 'name': PowerOnVM_Task, 'duration_secs': 0.649328} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.176617] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 780.180189] env[62385]: DEBUG nova.compute.manager [None req-56ce2e84-eef9-4802-abe7-c25f6074268e tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 780.181119] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a27a9de-64ef-4596-91eb-e6e09fd13077 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.227386] env[62385]: DEBUG nova.compute.manager [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 780.245733] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205677, 'name': CreateVM_Task, 'duration_secs': 0.626524} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.247501] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 780.247501] env[62385]: DEBUG oslo_concurrency.lockutils [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.247501] env[62385]: DEBUG oslo_concurrency.lockutils [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.247732] env[62385]: DEBUG oslo_concurrency.lockutils [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 780.248325] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1d90ad1-0028-46c1-a1e8-60f2690b358a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.255883] env[62385]: DEBUG oslo_vmware.api [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 780.255883] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]528821b4-5422-54b6-8fe3-cf562f18175a" [ 780.255883] env[62385]: _type = "Task" [ 780.255883] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.263817] env[62385]: DEBUG nova.scheduler.client.report [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 780.270296] env[62385]: DEBUG nova.virt.hardware [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 780.270693] env[62385]: DEBUG nova.virt.hardware [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 780.270885] env[62385]: DEBUG nova.virt.hardware [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 780.271126] env[62385]: DEBUG nova.virt.hardware [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 780.271348] env[62385]: DEBUG nova.virt.hardware [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 780.275025] env[62385]: DEBUG nova.virt.hardware [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 780.275025] env[62385]: DEBUG nova.virt.hardware [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 780.275025] env[62385]: DEBUG nova.virt.hardware [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 780.275025] env[62385]: DEBUG nova.virt.hardware [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 780.275025] env[62385]: DEBUG nova.virt.hardware [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 780.275289] env[62385]: DEBUG nova.virt.hardware [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 780.275289] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d31850b3-7f3c-4e79-acbc-2cb8e47ccd42 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.281730] env[62385]: DEBUG oslo_vmware.api [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]528821b4-5422-54b6-8fe3-cf562f18175a, 'name': SearchDatastore_Task, 'duration_secs': 0.019622} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.283160] env[62385]: DEBUG oslo_concurrency.lockutils [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.283468] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 780.283780] env[62385]: DEBUG oslo_concurrency.lockutils [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.283982] env[62385]: DEBUG oslo_concurrency.lockutils [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.284232] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 780.284946] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9c85d5e0-39e8-44c3-ad89-c21e866570f0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.291480] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71447337-63d1-419d-8f1f-863a43c2d639 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.298377] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 780.298641] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 780.300159] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c6ee3c6-cb60-4f3a-a557-c6ccdc7fc2ee {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.318840] env[62385]: DEBUG oslo_vmware.api [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 780.318840] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5208d371-7710-253a-9b81-1d32e5614a4f" [ 780.318840] env[62385]: _type = "Task" [ 780.318840] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.328132] env[62385]: DEBUG oslo_vmware.api [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5208d371-7710-253a-9b81-1d32e5614a4f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.360310] env[62385]: INFO nova.compute.manager [-] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Took 1.83 seconds to deallocate network for instance. [ 780.378976] env[62385]: DEBUG oslo_vmware.api [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205672, 'name': CloneVM_Task, 'duration_secs': 1.338727} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.379285] env[62385]: INFO nova.virt.vmwareapi.vmops [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Created linked-clone VM from snapshot [ 780.380098] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd0b7f4a-0d32-400a-b49c-189d439f1114 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.388324] env[62385]: DEBUG nova.virt.vmwareapi.images [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Uploading image 682eb0ec-9dc8-46f4-9a39-28fdbc76a00a {{(pid=62385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 780.416602] env[62385]: DEBUG oslo_vmware.rw_handles [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 780.416602] env[62385]: value = "vm-261172" [ 780.416602] env[62385]: _type = "VirtualMachine" [ 780.416602] env[62385]: }. {{(pid=62385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 780.417241] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-db38c9a1-8d15-41de-a813-735aa3aabf96 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.425663] env[62385]: DEBUG oslo_vmware.rw_handles [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lease: (returnval){ [ 780.425663] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52dce774-06e0-fd53-e861-8c4c36bb36da" [ 780.425663] env[62385]: _type = "HttpNfcLease" [ 780.425663] env[62385]: } obtained for exporting VM: (result){ [ 780.425663] env[62385]: value = "vm-261172" [ 780.425663] env[62385]: _type = "VirtualMachine" [ 780.425663] env[62385]: }. {{(pid=62385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 780.426029] env[62385]: DEBUG oslo_vmware.api [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the lease: (returnval){ [ 780.426029] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52dce774-06e0-fd53-e861-8c4c36bb36da" [ 780.426029] env[62385]: _type = "HttpNfcLease" [ 780.426029] env[62385]: } to be ready. {{(pid=62385) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 780.433188] env[62385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 780.433188] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52dce774-06e0-fd53-e861-8c4c36bb36da" [ 780.433188] env[62385]: _type = "HttpNfcLease" [ 780.433188] env[62385]: } is initializing. {{(pid=62385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 780.612246] env[62385]: DEBUG nova.compute.manager [req-caba3381-cee7-4987-95d1-e42cd7eeb177 req-67470d9a-8a87-4362-9ea2-0f8169fe2f46 service nova] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Received event network-vif-deleted-30e19299-99d0-4c98-8615-78f2f5f76dd5 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 780.632032] env[62385]: DEBUG nova.network.neutron [None req-eb8f459c-5746-44f6-b3e4-e73b2e5a968e tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 780.782025] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.615s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.782025] env[62385]: DEBUG nova.compute.manager [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 780.783054] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d7c0ede1-27a5-42bd-99d1-019b1c52ecae tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.586s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.783442] env[62385]: DEBUG nova.objects.instance [None req-d7c0ede1-27a5-42bd-99d1-019b1c52ecae tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Lazy-loading 'resources' on Instance uuid 749e0976-2e2b-4764-865c-2e630f2edbd1 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 780.831737] env[62385]: DEBUG oslo_vmware.api [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5208d371-7710-253a-9b81-1d32e5614a4f, 'name': SearchDatastore_Task, 'duration_secs': 0.01721} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.832856] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6cc1662c-9211-43c4-93ef-79667fe2436b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.839275] env[62385]: DEBUG oslo_vmware.api [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 780.839275] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]527a65a7-5317-7169-77a2-d2b4a4b9a4fb" [ 780.839275] env[62385]: _type = "Task" [ 780.839275] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.848782] env[62385]: DEBUG oslo_vmware.api [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]527a65a7-5317-7169-77a2-d2b4a4b9a4fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.869607] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b6753bc4-02bf-4300-903f-acd56a4146b9 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.936210] env[62385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 780.936210] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52dce774-06e0-fd53-e861-8c4c36bb36da" [ 780.936210] env[62385]: _type = "HttpNfcLease" [ 780.936210] env[62385]: } is ready. {{(pid=62385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 780.936587] env[62385]: DEBUG oslo_vmware.rw_handles [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 780.936587] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52dce774-06e0-fd53-e861-8c4c36bb36da" [ 780.936587] env[62385]: _type = "HttpNfcLease" [ 780.936587] env[62385]: }. {{(pid=62385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 780.937766] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c88089b-893c-4597-a34c-2aae8fff59a3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.945851] env[62385]: DEBUG oslo_vmware.rw_handles [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527c1b39-e62c-e7f3-a584-319f831f45e5/disk-0.vmdk from lease info. {{(pid=62385) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 780.946080] env[62385]: DEBUG oslo_vmware.rw_handles [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527c1b39-e62c-e7f3-a584-319f831f45e5/disk-0.vmdk for reading. {{(pid=62385) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 781.048764] env[62385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-7bfd7642-970c-4bd8-827f-3d9c2f34bb9b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.239293] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d08062cc-cece-446b-835d-8aef2e093696 tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Acquiring lock "f77dc5ab-7e63-4a1b-9c87-e474a7e28c93" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.239573] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d08062cc-cece-446b-835d-8aef2e093696 tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Lock "f77dc5ab-7e63-4a1b-9c87-e474a7e28c93" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.239776] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d08062cc-cece-446b-835d-8aef2e093696 tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Acquiring lock "f77dc5ab-7e63-4a1b-9c87-e474a7e28c93-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.239996] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d08062cc-cece-446b-835d-8aef2e093696 tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Lock "f77dc5ab-7e63-4a1b-9c87-e474a7e28c93-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.240287] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d08062cc-cece-446b-835d-8aef2e093696 tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Lock "f77dc5ab-7e63-4a1b-9c87-e474a7e28c93-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.243164] env[62385]: INFO nova.compute.manager [None req-d08062cc-cece-446b-835d-8aef2e093696 tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Terminating instance [ 781.247041] env[62385]: DEBUG nova.compute.manager [None req-d08062cc-cece-446b-835d-8aef2e093696 tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 781.247254] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d08062cc-cece-446b-835d-8aef2e093696 tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 781.248341] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38c37a44-570a-4615-9b6f-bf127b5402bb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.259587] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d08062cc-cece-446b-835d-8aef2e093696 tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 781.262056] env[62385]: DEBUG nova.virt.hardware [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 781.262578] env[62385]: DEBUG nova.virt.hardware [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 781.262899] env[62385]: DEBUG nova.virt.hardware [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 781.263248] env[62385]: DEBUG nova.virt.hardware [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 781.263582] env[62385]: DEBUG nova.virt.hardware [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 781.264197] env[62385]: DEBUG nova.virt.hardware [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 781.264197] env[62385]: DEBUG nova.virt.hardware [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 781.264376] env[62385]: DEBUG nova.virt.hardware [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 781.264749] env[62385]: DEBUG nova.virt.hardware [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 781.265083] env[62385]: DEBUG nova.virt.hardware [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 781.265366] env[62385]: DEBUG nova.virt.hardware [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 781.266094] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-75a0c108-9bd3-4c0c-aa51-3f2d476a034f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.269027] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d45c888c-3da9-4760-8107-7d3d85ef32a3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.280153] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ded042a8-0967-425a-b868-4c2b57489021 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.286774] env[62385]: DEBUG nova.compute.utils [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 781.288828] env[62385]: DEBUG oslo_vmware.api [None req-d08062cc-cece-446b-835d-8aef2e093696 tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Waiting for the task: (returnval){ [ 781.288828] env[62385]: value = "task-1205681" [ 781.288828] env[62385]: _type = "Task" [ 781.288828] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.289830] env[62385]: DEBUG nova.compute.manager [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 781.290010] env[62385]: DEBUG nova.network.neutron [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 781.308031] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Instance VIF info [] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 781.314047] env[62385]: DEBUG oslo.service.loopingcall [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 781.317925] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 781.319033] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f3ce11f8-b01b-45fa-8b41-17569702f20b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.338015] env[62385]: DEBUG oslo_vmware.api [None req-d08062cc-cece-446b-835d-8aef2e093696 tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Task: {'id': task-1205681, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.347306] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 781.347306] env[62385]: value = "task-1205682" [ 781.347306] env[62385]: _type = "Task" [ 781.347306] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.360212] env[62385]: DEBUG oslo_vmware.api [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]527a65a7-5317-7169-77a2-d2b4a4b9a4fb, 'name': SearchDatastore_Task, 'duration_secs': 0.037032} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.361127] env[62385]: DEBUG oslo_concurrency.lockutils [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.361466] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 0d823fcc-cbba-41f6-a47c-2f57d3e79948/0d823fcc-cbba-41f6-a47c-2f57d3e79948.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 781.361703] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4db6207a-5280-4bf8-91a5-de63aca916d0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.365037] env[62385]: DEBUG nova.policy [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'effe2cfd64d44e32a258588162812cba', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '83d39da6eb0541fea18f25acc9df3eef', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 781.370541] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205682, 'name': CreateVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.376950] env[62385]: DEBUG oslo_vmware.api [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 781.376950] env[62385]: value = "task-1205683" [ 781.376950] env[62385]: _type = "Task" [ 781.376950] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.386239] env[62385]: DEBUG oslo_vmware.api [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1205683, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.701207] env[62385]: DEBUG nova.network.neutron [None req-eb8f459c-5746-44f6-b3e4-e73b2e5a968e tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Updating instance_info_cache with network_info: [{"id": "fdae3362-d2d0-4a5e-80ff-c92e442ef9f4", "address": "fa:16:3e:87:99:cf", "network": {"id": "6a072c47-f3cd-4e88-a6f4-bb03318b678f", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1088193534-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.199", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ff114d055f440e0afc6c87e2ab6b4e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b5c60ce-845e-4506-bc10-348461fece6d", "external-id": "nsx-vlan-transportzone-831", "segmentation_id": 831, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfdae3362-d2", "ovs_interfaceid": "fdae3362-d2d0-4a5e-80ff-c92e442ef9f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.793032] env[62385]: DEBUG nova.compute.manager [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 781.818555] env[62385]: DEBUG oslo_vmware.api [None req-d08062cc-cece-446b-835d-8aef2e093696 tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Task: {'id': task-1205681, 'name': PowerOffVM_Task, 'duration_secs': 0.268185} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.818982] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d08062cc-cece-446b-835d-8aef2e093696 tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 781.819425] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d08062cc-cece-446b-835d-8aef2e093696 tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 781.820109] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d956da11-1cba-43ff-918e-e12ac6c3bace {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.838432] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4135db96-a3dc-4b00-a022-a9ef2cddc7ac {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.853526] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54d68173-4dba-42d9-93b4-16f96d887e05 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.902612] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be38fc89-fe94-4b2e-8ba5-7c24332fae22 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.906595] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205682, 'name': CreateVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.907634] env[62385]: DEBUG nova.network.neutron [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Successfully created port: c14a2370-3801-4b30-ad4f-e86a6b5c133d {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 781.922558] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c1bab34-0fc2-4c4b-990d-a59a4760cb5d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.927829] env[62385]: DEBUG oslo_vmware.api [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1205683, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.943829] env[62385]: DEBUG nova.compute.provider_tree [None req-d7c0ede1-27a5-42bd-99d1-019b1c52ecae tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 782.010219] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d08062cc-cece-446b-835d-8aef2e093696 tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 782.010606] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d08062cc-cece-446b-835d-8aef2e093696 tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Deleting contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 782.010825] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-d08062cc-cece-446b-835d-8aef2e093696 tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Deleting the datastore file [datastore2] f77dc5ab-7e63-4a1b-9c87-e474a7e28c93 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 782.011207] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dd046f38-1a17-4fd8-85c5-06dc60085f56 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.026512] env[62385]: DEBUG oslo_vmware.api [None req-d08062cc-cece-446b-835d-8aef2e093696 tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Waiting for the task: (returnval){ [ 782.026512] env[62385]: value = "task-1205685" [ 782.026512] env[62385]: _type = "Task" [ 782.026512] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.038066] env[62385]: DEBUG oslo_vmware.api [None req-d08062cc-cece-446b-835d-8aef2e093696 tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Task: {'id': task-1205685, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.113664] env[62385]: DEBUG nova.network.neutron [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Successfully updated port: 854fc77a-662c-4fbe-a30c-5726d1ed6088 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 782.203244] env[62385]: DEBUG oslo_concurrency.lockutils [None req-eb8f459c-5746-44f6-b3e4-e73b2e5a968e tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Releasing lock "refresh_cache-f50ddafe-f5a9-422b-b0f4-46c1b111dfbe" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.203529] env[62385]: DEBUG nova.compute.manager [None req-eb8f459c-5746-44f6-b3e4-e73b2e5a968e tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Inject network info {{(pid=62385) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7219}} [ 782.203803] env[62385]: DEBUG nova.compute.manager [None req-eb8f459c-5746-44f6-b3e4-e73b2e5a968e tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] network_info to inject: |[{"id": "fdae3362-d2d0-4a5e-80ff-c92e442ef9f4", "address": "fa:16:3e:87:99:cf", "network": {"id": "6a072c47-f3cd-4e88-a6f4-bb03318b678f", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1088193534-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.199", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ff114d055f440e0afc6c87e2ab6b4e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b5c60ce-845e-4506-bc10-348461fece6d", "external-id": "nsx-vlan-transportzone-831", "segmentation_id": 831, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfdae3362-d2", "ovs_interfaceid": "fdae3362-d2d0-4a5e-80ff-c92e442ef9f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7220}} [ 782.209318] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-eb8f459c-5746-44f6-b3e4-e73b2e5a968e tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Reconfiguring VM instance to set the machine id {{(pid=62385) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1802}} [ 782.209703] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2b605b9e-e61b-41d1-bb44-2196a151c196 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.230708] env[62385]: DEBUG oslo_vmware.api [None req-eb8f459c-5746-44f6-b3e4-e73b2e5a968e tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Waiting for the task: (returnval){ [ 782.230708] env[62385]: value = "task-1205686" [ 782.230708] env[62385]: _type = "Task" [ 782.230708] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.242790] env[62385]: DEBUG oslo_vmware.api [None req-eb8f459c-5746-44f6-b3e4-e73b2e5a968e tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Task: {'id': task-1205686, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.271532] env[62385]: DEBUG oslo_vmware.rw_handles [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52bc6476-0b00-6dd2-14c5-9b96ff5f7900/disk-0.vmdk. {{(pid=62385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 782.272452] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfc53079-34f7-4dd7-ae1d-f7e3673beb14 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.281502] env[62385]: DEBUG oslo_vmware.rw_handles [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52bc6476-0b00-6dd2-14c5-9b96ff5f7900/disk-0.vmdk is in state: ready. {{(pid=62385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 782.281718] env[62385]: ERROR oslo_vmware.rw_handles [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52bc6476-0b00-6dd2-14c5-9b96ff5f7900/disk-0.vmdk due to incomplete transfer. [ 782.281964] env[62385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-d9b105b9-7083-4c6c-a796-4f737e1c8824 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.290670] env[62385]: DEBUG oslo_vmware.rw_handles [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52bc6476-0b00-6dd2-14c5-9b96ff5f7900/disk-0.vmdk. {{(pid=62385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 782.290937] env[62385]: DEBUG nova.virt.vmwareapi.images [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Uploaded image e1482d72-7837-486f-b780-15965cbc5977 to the Glance image server {{(pid=62385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 782.293172] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Destroying the VM {{(pid=62385) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 782.293538] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-68b0ca4e-4c26-43da-8007-9947ff9fb5ff {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.302496] env[62385]: DEBUG oslo_vmware.api [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Waiting for the task: (returnval){ [ 782.302496] env[62385]: value = "task-1205687" [ 782.302496] env[62385]: _type = "Task" [ 782.302496] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.312490] env[62385]: DEBUG oslo_vmware.api [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205687, 'name': Destroy_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.362171] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205682, 'name': CreateVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.405458] env[62385]: DEBUG oslo_vmware.api [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1205683, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.613781} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.405869] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 0d823fcc-cbba-41f6-a47c-2f57d3e79948/0d823fcc-cbba-41f6-a47c-2f57d3e79948.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 782.406162] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 782.406477] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a3857f82-b342-4df7-8c53-d9f1941fa656 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.415056] env[62385]: DEBUG oslo_vmware.api [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 782.415056] env[62385]: value = "task-1205688" [ 782.415056] env[62385]: _type = "Task" [ 782.415056] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.426487] env[62385]: DEBUG oslo_vmware.api [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1205688, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.447395] env[62385]: DEBUG nova.scheduler.client.report [None req-d7c0ede1-27a5-42bd-99d1-019b1c52ecae tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 782.538286] env[62385]: DEBUG oslo_vmware.api [None req-d08062cc-cece-446b-835d-8aef2e093696 tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Task: {'id': task-1205685, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.284609} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.538593] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-d08062cc-cece-446b-835d-8aef2e093696 tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 782.538788] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d08062cc-cece-446b-835d-8aef2e093696 tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Deleted contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 782.538969] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d08062cc-cece-446b-835d-8aef2e093696 tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 782.539166] env[62385]: INFO nova.compute.manager [None req-d08062cc-cece-446b-835d-8aef2e093696 tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Took 1.29 seconds to destroy the instance on the hypervisor. [ 782.539417] env[62385]: DEBUG oslo.service.loopingcall [None req-d08062cc-cece-446b-835d-8aef2e093696 tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 782.539620] env[62385]: DEBUG nova.compute.manager [-] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 782.539701] env[62385]: DEBUG nova.network.neutron [-] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 782.616535] env[62385]: DEBUG oslo_concurrency.lockutils [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Acquiring lock "refresh_cache-b4970e1a-2b29-44e9-b79e-cda3c32dddcd" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.616722] env[62385]: DEBUG oslo_concurrency.lockutils [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Acquired lock "refresh_cache-b4970e1a-2b29-44e9-b79e-cda3c32dddcd" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.616932] env[62385]: DEBUG nova.network.neutron [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 782.741889] env[62385]: DEBUG oslo_vmware.api [None req-eb8f459c-5746-44f6-b3e4-e73b2e5a968e tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Task: {'id': task-1205686, 'name': ReconfigVM_Task, 'duration_secs': 0.192031} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.742205] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-eb8f459c-5746-44f6-b3e4-e73b2e5a968e tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Reconfigured VM instance to set the machine id {{(pid=62385) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1805}} [ 782.804899] env[62385]: DEBUG nova.compute.manager [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 782.817472] env[62385]: DEBUG oslo_vmware.api [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205687, 'name': Destroy_Task} progress is 33%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.826957] env[62385]: DEBUG nova.virt.hardware [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 782.827156] env[62385]: DEBUG nova.virt.hardware [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 782.827506] env[62385]: DEBUG nova.virt.hardware [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 782.827506] env[62385]: DEBUG nova.virt.hardware [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 782.827602] env[62385]: DEBUG nova.virt.hardware [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 782.827700] env[62385]: DEBUG nova.virt.hardware [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 782.827957] env[62385]: DEBUG nova.virt.hardware [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 782.828417] env[62385]: DEBUG nova.virt.hardware [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 782.828684] env[62385]: DEBUG nova.virt.hardware [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 782.828899] env[62385]: DEBUG nova.virt.hardware [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 782.829132] env[62385]: DEBUG nova.virt.hardware [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 782.830123] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09ac17e1-3d6b-4594-9ab2-db77e432c2f5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.840685] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1de91c74-63f2-476c-9169-41336464686e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.862453] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205682, 'name': CreateVM_Task, 'duration_secs': 1.382063} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.862630] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 782.863076] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.863244] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.863591] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 782.863856] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-05ecf2fe-fb35-4c3f-834e-ebb89e8cf9f0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.869913] env[62385]: DEBUG oslo_vmware.api [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Waiting for the task: (returnval){ [ 782.869913] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52385e30-8b5c-2aea-2242-bf0f6c7a42f7" [ 782.869913] env[62385]: _type = "Task" [ 782.869913] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.879118] env[62385]: DEBUG oslo_vmware.api [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52385e30-8b5c-2aea-2242-bf0f6c7a42f7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.894589] env[62385]: DEBUG nova.compute.manager [req-4ea14bd8-58f7-4855-aef5-dc93f595f1b2 req-d081a08a-051f-46a5-abdd-b042810fcfa9 service nova] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Received event network-changed-fdae3362-d2d0-4a5e-80ff-c92e442ef9f4 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 782.894765] env[62385]: DEBUG nova.compute.manager [req-4ea14bd8-58f7-4855-aef5-dc93f595f1b2 req-d081a08a-051f-46a5-abdd-b042810fcfa9 service nova] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Refreshing instance network info cache due to event network-changed-fdae3362-d2d0-4a5e-80ff-c92e442ef9f4. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 782.894991] env[62385]: DEBUG oslo_concurrency.lockutils [req-4ea14bd8-58f7-4855-aef5-dc93f595f1b2 req-d081a08a-051f-46a5-abdd-b042810fcfa9 service nova] Acquiring lock "refresh_cache-f50ddafe-f5a9-422b-b0f4-46c1b111dfbe" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.895206] env[62385]: DEBUG oslo_concurrency.lockutils [req-4ea14bd8-58f7-4855-aef5-dc93f595f1b2 req-d081a08a-051f-46a5-abdd-b042810fcfa9 service nova] Acquired lock "refresh_cache-f50ddafe-f5a9-422b-b0f4-46c1b111dfbe" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.895381] env[62385]: DEBUG nova.network.neutron [req-4ea14bd8-58f7-4855-aef5-dc93f595f1b2 req-d081a08a-051f-46a5-abdd-b042810fcfa9 service nova] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Refreshing network info cache for port fdae3362-d2d0-4a5e-80ff-c92e442ef9f4 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 782.929029] env[62385]: DEBUG oslo_vmware.api [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1205688, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080233} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.929029] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 782.929029] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7859c34d-0855-4538-8b7b-772de69c364f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.954902] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Reconfiguring VM instance instance-0000003b to attach disk [datastore1] 0d823fcc-cbba-41f6-a47c-2f57d3e79948/0d823fcc-cbba-41f6-a47c-2f57d3e79948.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 782.956158] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d7c0ede1-27a5-42bd-99d1-019b1c52ecae tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.174s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.958637] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8d65acde-242c-4226-b8eb-43bdf4528c1b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.975583] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9e6a2dab-7bf8-4fe5-a0e1-11e4d30c5c85 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.768s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.976041] env[62385]: DEBUG nova.objects.instance [None req-9e6a2dab-7bf8-4fe5-a0e1-11e4d30c5c85 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Lazy-loading 'resources' on Instance uuid e97ff020-61f3-4947-bb82-5c039ad17747 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 782.988386] env[62385]: DEBUG oslo_vmware.api [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 782.988386] env[62385]: value = "task-1205689" [ 782.988386] env[62385]: _type = "Task" [ 782.988386] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.999223] env[62385]: DEBUG oslo_vmware.api [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1205689, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.000888] env[62385]: INFO nova.scheduler.client.report [None req-d7c0ede1-27a5-42bd-99d1-019b1c52ecae tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Deleted allocations for instance 749e0976-2e2b-4764-865c-2e630f2edbd1 [ 783.175201] env[62385]: DEBUG nova.network.neutron [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 783.317899] env[62385]: DEBUG oslo_vmware.api [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205687, 'name': Destroy_Task, 'duration_secs': 0.605358} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.321018] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Destroyed the VM [ 783.321105] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Deleting Snapshot of the VM instance {{(pid=62385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 783.321870] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-8c64391b-5343-4ae7-a7be-151512f4e39f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.330888] env[62385]: DEBUG oslo_vmware.api [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Waiting for the task: (returnval){ [ 783.330888] env[62385]: value = "task-1205690" [ 783.330888] env[62385]: _type = "Task" [ 783.330888] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.340210] env[62385]: DEBUG oslo_vmware.api [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205690, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.382239] env[62385]: DEBUG oslo_vmware.api [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52385e30-8b5c-2aea-2242-bf0f6c7a42f7, 'name': SearchDatastore_Task, 'duration_secs': 0.016711} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.382610] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.382892] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 783.383275] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.383467] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.383684] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 783.383989] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0c2762cf-679f-4dc7-96d8-5b11262bbfbc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.388044] env[62385]: DEBUG nova.network.neutron [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Updating instance_info_cache with network_info: [{"id": "854fc77a-662c-4fbe-a30c-5726d1ed6088", "address": "fa:16:3e:a6:e0:c0", "network": {"id": "b563b70c-d3f1-4f82-97e9-defaf3f9d14b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-55206131-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "8eaa81326f084441a87ac05796958abb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa", "external-id": "nsx-vlan-transportzone-132", "segmentation_id": 132, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap854fc77a-66", "ovs_interfaceid": "854fc77a-662c-4fbe-a30c-5726d1ed6088", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.394442] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 783.394666] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 783.395474] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f269588-126b-4c1b-bed4-961ccb07fa1b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.406160] env[62385]: DEBUG oslo_vmware.api [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Waiting for the task: (returnval){ [ 783.406160] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52548f6e-dd2c-c044-c093-543771fd148d" [ 783.406160] env[62385]: _type = "Task" [ 783.406160] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.415598] env[62385]: DEBUG oslo_vmware.api [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52548f6e-dd2c-c044-c093-543771fd148d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.499588] env[62385]: DEBUG oslo_vmware.api [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1205689, 'name': ReconfigVM_Task, 'duration_secs': 0.511863} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.499882] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Reconfigured VM instance instance-0000003b to attach disk [datastore1] 0d823fcc-cbba-41f6-a47c-2f57d3e79948/0d823fcc-cbba-41f6-a47c-2f57d3e79948.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 783.500630] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-280b3739-e38e-4edb-ac73-3892c89df58a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.511500] env[62385]: DEBUG oslo_vmware.api [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 783.511500] env[62385]: value = "task-1205691" [ 783.511500] env[62385]: _type = "Task" [ 783.511500] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.511500] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d7c0ede1-27a5-42bd-99d1-019b1c52ecae tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Lock "749e0976-2e2b-4764-865c-2e630f2edbd1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.275s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.520989] env[62385]: DEBUG oslo_vmware.api [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1205691, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.548978] env[62385]: DEBUG nova.network.neutron [-] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.554757] env[62385]: DEBUG nova.network.neutron [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Successfully updated port: c14a2370-3801-4b30-ad4f-e86a6b5c133d {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 783.843171] env[62385]: DEBUG oslo_vmware.api [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205690, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.850187] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ebabc504-fb40-4408-abaf-177b5b541c69 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Acquiring lock "f50ddafe-f5a9-422b-b0f4-46c1b111dfbe" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.850535] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ebabc504-fb40-4408-abaf-177b5b541c69 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Lock "f50ddafe-f5a9-422b-b0f4-46c1b111dfbe" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.850762] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ebabc504-fb40-4408-abaf-177b5b541c69 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Acquiring lock "f50ddafe-f5a9-422b-b0f4-46c1b111dfbe-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.850952] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ebabc504-fb40-4408-abaf-177b5b541c69 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Lock "f50ddafe-f5a9-422b-b0f4-46c1b111dfbe-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.851148] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ebabc504-fb40-4408-abaf-177b5b541c69 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Lock "f50ddafe-f5a9-422b-b0f4-46c1b111dfbe-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.853598] env[62385]: INFO nova.compute.manager [None req-ebabc504-fb40-4408-abaf-177b5b541c69 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Terminating instance [ 783.855593] env[62385]: DEBUG nova.compute.manager [None req-ebabc504-fb40-4408-abaf-177b5b541c69 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 783.855761] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ebabc504-fb40-4408-abaf-177b5b541c69 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 783.856857] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4e6fcd4-4680-43fe-9ea4-80c4c75ab0bd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.870395] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebabc504-fb40-4408-abaf-177b5b541c69 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 783.870700] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8e42532f-da8d-4fd1-8518-ac0ba88e018c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.879381] env[62385]: DEBUG oslo_vmware.api [None req-ebabc504-fb40-4408-abaf-177b5b541c69 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Waiting for the task: (returnval){ [ 783.879381] env[62385]: value = "task-1205692" [ 783.879381] env[62385]: _type = "Task" [ 783.879381] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.896699] env[62385]: DEBUG oslo_concurrency.lockutils [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Releasing lock "refresh_cache-b4970e1a-2b29-44e9-b79e-cda3c32dddcd" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.897062] env[62385]: DEBUG nova.compute.manager [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Instance network_info: |[{"id": "854fc77a-662c-4fbe-a30c-5726d1ed6088", "address": "fa:16:3e:a6:e0:c0", "network": {"id": "b563b70c-d3f1-4f82-97e9-defaf3f9d14b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-55206131-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "8eaa81326f084441a87ac05796958abb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa", "external-id": "nsx-vlan-transportzone-132", "segmentation_id": 132, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap854fc77a-66", "ovs_interfaceid": "854fc77a-662c-4fbe-a30c-5726d1ed6088", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 783.897407] env[62385]: DEBUG oslo_vmware.api [None req-ebabc504-fb40-4408-abaf-177b5b541c69 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Task: {'id': task-1205692, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.900679] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a6:e0:c0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '854fc77a-662c-4fbe-a30c-5726d1ed6088', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 783.909016] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Creating folder: Project (8eaa81326f084441a87ac05796958abb). Parent ref: group-v261107. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 783.909893] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fb8e4a3a-64b7-49f4-ad9c-561836aa6f29 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.921986] env[62385]: DEBUG oslo_vmware.api [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52548f6e-dd2c-c044-c093-543771fd148d, 'name': SearchDatastore_Task, 'duration_secs': 0.016431} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.922917] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e93c4088-2c3e-4f1a-970a-ddd0ce51b9c7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.929369] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Created folder: Project (8eaa81326f084441a87ac05796958abb) in parent group-v261107. [ 783.929687] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Creating folder: Instances. Parent ref: group-v261177. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 783.930527] env[62385]: DEBUG nova.network.neutron [req-4ea14bd8-58f7-4855-aef5-dc93f595f1b2 req-d081a08a-051f-46a5-abdd-b042810fcfa9 service nova] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Updated VIF entry in instance network info cache for port fdae3362-d2d0-4a5e-80ff-c92e442ef9f4. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 783.930877] env[62385]: DEBUG nova.network.neutron [req-4ea14bd8-58f7-4855-aef5-dc93f595f1b2 req-d081a08a-051f-46a5-abdd-b042810fcfa9 service nova] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Updating instance_info_cache with network_info: [{"id": "fdae3362-d2d0-4a5e-80ff-c92e442ef9f4", "address": "fa:16:3e:87:99:cf", "network": {"id": "6a072c47-f3cd-4e88-a6f4-bb03318b678f", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1088193534-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.199", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ff114d055f440e0afc6c87e2ab6b4e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b5c60ce-845e-4506-bc10-348461fece6d", "external-id": "nsx-vlan-transportzone-831", "segmentation_id": 831, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfdae3362-d2", "ovs_interfaceid": "fdae3362-d2d0-4a5e-80ff-c92e442ef9f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.932694] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d02c8245-b22a-4a6a-88c3-f402032b576f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.936044] env[62385]: DEBUG oslo_vmware.api [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Waiting for the task: (returnval){ [ 783.936044] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52b52dcc-d168-077e-4021-1d4d0154bfc9" [ 783.936044] env[62385]: _type = "Task" [ 783.936044] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.946650] env[62385]: DEBUG oslo_vmware.api [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52b52dcc-d168-077e-4021-1d4d0154bfc9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.958988] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Created folder: Instances in parent group-v261177. [ 783.958988] env[62385]: DEBUG oslo.service.loopingcall [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 783.958988] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 783.958988] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-86463707-cf2e-4701-b455-7d7084b4f420 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.976504] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0be2e375-5ddd-4c73-9f06-22e6ffc8c703 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.986072] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36121c4d-0434-4af6-b6fa-881f65fb7c05 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.989826] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 783.989826] env[62385]: value = "task-1205695" [ 783.989826] env[62385]: _type = "Task" [ 783.989826] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.024588] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74509a8a-5d6e-472c-af63-e995704dc94d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.031159] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205695, 'name': CreateVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.037533] env[62385]: DEBUG oslo_vmware.api [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1205691, 'name': Rename_Task, 'duration_secs': 0.234665} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.040126] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 784.040673] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-69eb1d73-d2e5-45a7-b477-e4aa4a31cb39 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.043568] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf0eb7f8-909b-4386-8cd1-17b98a785429 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.051752] env[62385]: INFO nova.compute.manager [-] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Took 1.51 seconds to deallocate network for instance. [ 784.058404] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Acquiring lock "refresh_cache-fd7a77e2-6d76-4d91-bdef-e30333247aa9" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.058791] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Acquired lock "refresh_cache-fd7a77e2-6d76-4d91-bdef-e30333247aa9" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.058791] env[62385]: DEBUG nova.network.neutron [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 784.072358] env[62385]: DEBUG oslo_vmware.api [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 784.072358] env[62385]: value = "task-1205696" [ 784.072358] env[62385]: _type = "Task" [ 784.072358] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.072358] env[62385]: DEBUG nova.compute.provider_tree [None req-9e6a2dab-7bf8-4fe5-a0e1-11e4d30c5c85 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 784.084245] env[62385]: DEBUG oslo_vmware.api [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1205696, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.341705] env[62385]: DEBUG oslo_vmware.api [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205690, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.393720] env[62385]: DEBUG oslo_vmware.api [None req-ebabc504-fb40-4408-abaf-177b5b541c69 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Task: {'id': task-1205692, 'name': PowerOffVM_Task, 'duration_secs': 0.219321} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.394108] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebabc504-fb40-4408-abaf-177b5b541c69 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 784.394219] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ebabc504-fb40-4408-abaf-177b5b541c69 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 784.394515] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6a635200-464d-42ae-95ad-6202a52e27ab {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.435351] env[62385]: DEBUG oslo_concurrency.lockutils [req-4ea14bd8-58f7-4855-aef5-dc93f595f1b2 req-d081a08a-051f-46a5-abdd-b042810fcfa9 service nova] Releasing lock "refresh_cache-f50ddafe-f5a9-422b-b0f4-46c1b111dfbe" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.435642] env[62385]: DEBUG nova.compute.manager [req-4ea14bd8-58f7-4855-aef5-dc93f595f1b2 req-d081a08a-051f-46a5-abdd-b042810fcfa9 service nova] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Received event network-vif-plugged-854fc77a-662c-4fbe-a30c-5726d1ed6088 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 784.435852] env[62385]: DEBUG oslo_concurrency.lockutils [req-4ea14bd8-58f7-4855-aef5-dc93f595f1b2 req-d081a08a-051f-46a5-abdd-b042810fcfa9 service nova] Acquiring lock "b4970e1a-2b29-44e9-b79e-cda3c32dddcd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.436298] env[62385]: DEBUG oslo_concurrency.lockutils [req-4ea14bd8-58f7-4855-aef5-dc93f595f1b2 req-d081a08a-051f-46a5-abdd-b042810fcfa9 service nova] Lock "b4970e1a-2b29-44e9-b79e-cda3c32dddcd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.436498] env[62385]: DEBUG oslo_concurrency.lockutils [req-4ea14bd8-58f7-4855-aef5-dc93f595f1b2 req-d081a08a-051f-46a5-abdd-b042810fcfa9 service nova] Lock "b4970e1a-2b29-44e9-b79e-cda3c32dddcd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.436722] env[62385]: DEBUG nova.compute.manager [req-4ea14bd8-58f7-4855-aef5-dc93f595f1b2 req-d081a08a-051f-46a5-abdd-b042810fcfa9 service nova] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] No waiting events found dispatching network-vif-plugged-854fc77a-662c-4fbe-a30c-5726d1ed6088 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 784.436934] env[62385]: WARNING nova.compute.manager [req-4ea14bd8-58f7-4855-aef5-dc93f595f1b2 req-d081a08a-051f-46a5-abdd-b042810fcfa9 service nova] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Received unexpected event network-vif-plugged-854fc77a-662c-4fbe-a30c-5726d1ed6088 for instance with vm_state building and task_state spawning. [ 784.437152] env[62385]: DEBUG nova.compute.manager [req-4ea14bd8-58f7-4855-aef5-dc93f595f1b2 req-d081a08a-051f-46a5-abdd-b042810fcfa9 service nova] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Received event network-changed-854fc77a-662c-4fbe-a30c-5726d1ed6088 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 784.437351] env[62385]: DEBUG nova.compute.manager [req-4ea14bd8-58f7-4855-aef5-dc93f595f1b2 req-d081a08a-051f-46a5-abdd-b042810fcfa9 service nova] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Refreshing instance network info cache due to event network-changed-854fc77a-662c-4fbe-a30c-5726d1ed6088. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 784.437573] env[62385]: DEBUG oslo_concurrency.lockutils [req-4ea14bd8-58f7-4855-aef5-dc93f595f1b2 req-d081a08a-051f-46a5-abdd-b042810fcfa9 service nova] Acquiring lock "refresh_cache-b4970e1a-2b29-44e9-b79e-cda3c32dddcd" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.437745] env[62385]: DEBUG oslo_concurrency.lockutils [req-4ea14bd8-58f7-4855-aef5-dc93f595f1b2 req-d081a08a-051f-46a5-abdd-b042810fcfa9 service nova] Acquired lock "refresh_cache-b4970e1a-2b29-44e9-b79e-cda3c32dddcd" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.437945] env[62385]: DEBUG nova.network.neutron [req-4ea14bd8-58f7-4855-aef5-dc93f595f1b2 req-d081a08a-051f-46a5-abdd-b042810fcfa9 service nova] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Refreshing network info cache for port 854fc77a-662c-4fbe-a30c-5726d1ed6088 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 784.450109] env[62385]: DEBUG oslo_vmware.api [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52b52dcc-d168-077e-4021-1d4d0154bfc9, 'name': SearchDatastore_Task, 'duration_secs': 0.015287} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.451116] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.451543] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7/8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 784.451844] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2f6b1f57-ed0c-4222-bee7-16b96a13323b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.462131] env[62385]: DEBUG oslo_vmware.api [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Waiting for the task: (returnval){ [ 784.462131] env[62385]: value = "task-1205698" [ 784.462131] env[62385]: _type = "Task" [ 784.462131] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.472798] env[62385]: DEBUG oslo_vmware.api [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': task-1205698, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.501331] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205695, 'name': CreateVM_Task} progress is 25%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.503923] env[62385]: DEBUG nova.compute.manager [req-435f8ee0-bc96-4f54-acc6-718d2c593f57 req-60cd46a9-9318-4e9b-ac49-6d6a3d7275f8 service nova] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Received event network-vif-plugged-c14a2370-3801-4b30-ad4f-e86a6b5c133d {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 784.504178] env[62385]: DEBUG oslo_concurrency.lockutils [req-435f8ee0-bc96-4f54-acc6-718d2c593f57 req-60cd46a9-9318-4e9b-ac49-6d6a3d7275f8 service nova] Acquiring lock "fd7a77e2-6d76-4d91-bdef-e30333247aa9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.504402] env[62385]: DEBUG oslo_concurrency.lockutils [req-435f8ee0-bc96-4f54-acc6-718d2c593f57 req-60cd46a9-9318-4e9b-ac49-6d6a3d7275f8 service nova] Lock "fd7a77e2-6d76-4d91-bdef-e30333247aa9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.504579] env[62385]: DEBUG oslo_concurrency.lockutils [req-435f8ee0-bc96-4f54-acc6-718d2c593f57 req-60cd46a9-9318-4e9b-ac49-6d6a3d7275f8 service nova] Lock "fd7a77e2-6d76-4d91-bdef-e30333247aa9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.504802] env[62385]: DEBUG nova.compute.manager [req-435f8ee0-bc96-4f54-acc6-718d2c593f57 req-60cd46a9-9318-4e9b-ac49-6d6a3d7275f8 service nova] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] No waiting events found dispatching network-vif-plugged-c14a2370-3801-4b30-ad4f-e86a6b5c133d {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 784.504949] env[62385]: WARNING nova.compute.manager [req-435f8ee0-bc96-4f54-acc6-718d2c593f57 req-60cd46a9-9318-4e9b-ac49-6d6a3d7275f8 service nova] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Received unexpected event network-vif-plugged-c14a2370-3801-4b30-ad4f-e86a6b5c133d for instance with vm_state building and task_state spawning. [ 784.561058] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d08062cc-cece-446b-835d-8aef2e093696 tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.576863] env[62385]: DEBUG nova.scheduler.client.report [None req-9e6a2dab-7bf8-4fe5-a0e1-11e4d30c5c85 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 784.592323] env[62385]: DEBUG oslo_vmware.api [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1205696, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.613938] env[62385]: DEBUG nova.network.neutron [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 784.634898] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ebabc504-fb40-4408-abaf-177b5b541c69 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 784.635186] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ebabc504-fb40-4408-abaf-177b5b541c69 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Deleting contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 784.635440] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ebabc504-fb40-4408-abaf-177b5b541c69 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Deleting the datastore file [datastore2] f50ddafe-f5a9-422b-b0f4-46c1b111dfbe {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 784.636427] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d53ba782-d684-490e-832b-bb46f0bc09b3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.645692] env[62385]: DEBUG oslo_vmware.api [None req-ebabc504-fb40-4408-abaf-177b5b541c69 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Waiting for the task: (returnval){ [ 784.645692] env[62385]: value = "task-1205699" [ 784.645692] env[62385]: _type = "Task" [ 784.645692] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.659837] env[62385]: DEBUG oslo_vmware.api [None req-ebabc504-fb40-4408-abaf-177b5b541c69 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Task: {'id': task-1205699, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.851581] env[62385]: DEBUG oslo_vmware.api [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205690, 'name': RemoveSnapshot_Task, 'duration_secs': 1.106178} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.851901] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Deleted Snapshot of the VM instance {{(pid=62385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 784.852248] env[62385]: INFO nova.compute.manager [None req-08524f2e-4525-4b7e-a172-0bf0af886cd6 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Took 17.00 seconds to snapshot the instance on the hypervisor. [ 784.926628] env[62385]: DEBUG nova.network.neutron [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Updating instance_info_cache with network_info: [{"id": "c14a2370-3801-4b30-ad4f-e86a6b5c133d", "address": "fa:16:3e:77:22:db", "network": {"id": "ae13fe4b-9c18-45eb-8619-c58d283abd3f", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-432568731-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83d39da6eb0541fea18f25acc9df3eef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc14a2370-38", "ovs_interfaceid": "c14a2370-3801-4b30-ad4f-e86a6b5c133d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.984941] env[62385]: DEBUG oslo_vmware.api [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': task-1205698, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.003706] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205695, 'name': CreateVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.087501] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9e6a2dab-7bf8-4fe5-a0e1-11e4d30c5c85 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.112s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.089817] env[62385]: DEBUG oslo_vmware.api [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1205696, 'name': PowerOnVM_Task, 'duration_secs': 0.791994} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.090373] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 32.106s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.090658] env[62385]: DEBUG nova.objects.instance [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62385) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 785.093838] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 785.094077] env[62385]: INFO nova.compute.manager [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Took 9.73 seconds to spawn the instance on the hypervisor. [ 785.094443] env[62385]: DEBUG nova.compute.manager [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 785.095581] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-246246ca-984f-472c-884e-cc3f87cfa94a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.118272] env[62385]: INFO nova.scheduler.client.report [None req-9e6a2dab-7bf8-4fe5-a0e1-11e4d30c5c85 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Deleted allocations for instance e97ff020-61f3-4947-bb82-5c039ad17747 [ 785.161772] env[62385]: DEBUG oslo_vmware.api [None req-ebabc504-fb40-4408-abaf-177b5b541c69 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Task: {'id': task-1205699, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.311454} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.162156] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ebabc504-fb40-4408-abaf-177b5b541c69 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 785.162391] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ebabc504-fb40-4408-abaf-177b5b541c69 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Deleted contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 785.162643] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ebabc504-fb40-4408-abaf-177b5b541c69 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 785.162864] env[62385]: INFO nova.compute.manager [None req-ebabc504-fb40-4408-abaf-177b5b541c69 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Took 1.31 seconds to destroy the instance on the hypervisor. [ 785.163209] env[62385]: DEBUG oslo.service.loopingcall [None req-ebabc504-fb40-4408-abaf-177b5b541c69 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 785.163730] env[62385]: DEBUG nova.compute.manager [-] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 785.163874] env[62385]: DEBUG nova.network.neutron [-] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 785.410528] env[62385]: DEBUG nova.network.neutron [req-4ea14bd8-58f7-4855-aef5-dc93f595f1b2 req-d081a08a-051f-46a5-abdd-b042810fcfa9 service nova] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Updated VIF entry in instance network info cache for port 854fc77a-662c-4fbe-a30c-5726d1ed6088. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 785.410528] env[62385]: DEBUG nova.network.neutron [req-4ea14bd8-58f7-4855-aef5-dc93f595f1b2 req-d081a08a-051f-46a5-abdd-b042810fcfa9 service nova] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Updating instance_info_cache with network_info: [{"id": "854fc77a-662c-4fbe-a30c-5726d1ed6088", "address": "fa:16:3e:a6:e0:c0", "network": {"id": "b563b70c-d3f1-4f82-97e9-defaf3f9d14b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-55206131-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "8eaa81326f084441a87ac05796958abb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa", "external-id": "nsx-vlan-transportzone-132", "segmentation_id": 132, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap854fc77a-66", "ovs_interfaceid": "854fc77a-662c-4fbe-a30c-5726d1ed6088", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.429825] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Releasing lock "refresh_cache-fd7a77e2-6d76-4d91-bdef-e30333247aa9" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.430187] env[62385]: DEBUG nova.compute.manager [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Instance network_info: |[{"id": "c14a2370-3801-4b30-ad4f-e86a6b5c133d", "address": "fa:16:3e:77:22:db", "network": {"id": "ae13fe4b-9c18-45eb-8619-c58d283abd3f", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-432568731-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83d39da6eb0541fea18f25acc9df3eef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc14a2370-38", "ovs_interfaceid": "c14a2370-3801-4b30-ad4f-e86a6b5c133d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 785.430644] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:77:22:db', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3c29724c-5452-441a-8060-5bf89d1f5847', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c14a2370-3801-4b30-ad4f-e86a6b5c133d', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 785.439010] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Creating folder: Project (83d39da6eb0541fea18f25acc9df3eef). Parent ref: group-v261107. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 785.439604] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4a0eaa45-a8e3-4c54-851c-70ea84b2f6cb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.456821] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Created folder: Project (83d39da6eb0541fea18f25acc9df3eef) in parent group-v261107. [ 785.457077] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Creating folder: Instances. Parent ref: group-v261180. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 785.457354] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3146738b-65e4-47ed-91fa-f8f55cbe5b94 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.470945] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Created folder: Instances in parent group-v261180. [ 785.471259] env[62385]: DEBUG oslo.service.loopingcall [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 785.471914] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 785.472097] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-42b4d167-dd7c-4cae-9c04-66e9e98eb142 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.490101] env[62385]: DEBUG oslo_vmware.api [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': task-1205698, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.64752} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.490826] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7/8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 785.491076] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 785.491355] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a4db3193-e26e-4eb6-aa63-74f92421fbbb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.497779] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 785.497779] env[62385]: value = "task-1205702" [ 785.497779] env[62385]: _type = "Task" [ 785.497779] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.507040] env[62385]: DEBUG oslo_vmware.api [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Waiting for the task: (returnval){ [ 785.507040] env[62385]: value = "task-1205703" [ 785.507040] env[62385]: _type = "Task" [ 785.507040] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.507431] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205695, 'name': CreateVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.513855] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205702, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.519391] env[62385]: DEBUG oslo_vmware.api [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': task-1205703, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.617278] env[62385]: INFO nova.compute.manager [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Took 41.10 seconds to build instance. [ 785.627387] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9e6a2dab-7bf8-4fe5-a0e1-11e4d30c5c85 tempest-MultipleCreateTestJSON-1216687210 tempest-MultipleCreateTestJSON-1216687210-project-member] Lock "e97ff020-61f3-4947-bb82-5c039ad17747" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.504s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.815199] env[62385]: DEBUG nova.compute.manager [req-9903f7e5-b9a9-4541-8ec2-7ccb64a26e2c req-09ede329-8e9a-4286-b290-3df1ad59d89a service nova] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Received event network-vif-deleted-326dbe5f-be23-4610-bc79-079259f3a2c3 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 785.815477] env[62385]: DEBUG nova.compute.manager [req-9903f7e5-b9a9-4541-8ec2-7ccb64a26e2c req-09ede329-8e9a-4286-b290-3df1ad59d89a service nova] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Received event network-changed-b838271e-07e8-4880-980a-703c1bbd4da5 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 785.815846] env[62385]: DEBUG nova.compute.manager [req-9903f7e5-b9a9-4541-8ec2-7ccb64a26e2c req-09ede329-8e9a-4286-b290-3df1ad59d89a service nova] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Refreshing instance network info cache due to event network-changed-b838271e-07e8-4880-980a-703c1bbd4da5. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 785.816265] env[62385]: DEBUG oslo_concurrency.lockutils [req-9903f7e5-b9a9-4541-8ec2-7ccb64a26e2c req-09ede329-8e9a-4286-b290-3df1ad59d89a service nova] Acquiring lock "refresh_cache-65bb1d14-84b2-4e75-acdc-dc674a035101" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 785.816535] env[62385]: DEBUG oslo_concurrency.lockutils [req-9903f7e5-b9a9-4541-8ec2-7ccb64a26e2c req-09ede329-8e9a-4286-b290-3df1ad59d89a service nova] Acquired lock "refresh_cache-65bb1d14-84b2-4e75-acdc-dc674a035101" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.817098] env[62385]: DEBUG nova.network.neutron [req-9903f7e5-b9a9-4541-8ec2-7ccb64a26e2c req-09ede329-8e9a-4286-b290-3df1ad59d89a service nova] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Refreshing network info cache for port b838271e-07e8-4880-980a-703c1bbd4da5 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 785.913089] env[62385]: DEBUG oslo_concurrency.lockutils [req-4ea14bd8-58f7-4855-aef5-dc93f595f1b2 req-d081a08a-051f-46a5-abdd-b042810fcfa9 service nova] Releasing lock "refresh_cache-b4970e1a-2b29-44e9-b79e-cda3c32dddcd" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 786.010431] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205695, 'name': CreateVM_Task, 'duration_secs': 1.82901} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.016771] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 786.017083] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205702, 'name': CreateVM_Task, 'duration_secs': 0.510834} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.017780] env[62385]: DEBUG oslo_concurrency.lockutils [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.017972] env[62385]: DEBUG oslo_concurrency.lockutils [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.018381] env[62385]: DEBUG oslo_concurrency.lockutils [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 786.018591] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 786.019265] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ca93cb9-f363-4625-85a2-9195d0c22a15 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.021499] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.024792] env[62385]: DEBUG oslo_vmware.api [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': task-1205703, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086872} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.025477] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 786.027737] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d9c6ee7-ab79-4f50-99bd-058122d11b69 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.030677] env[62385]: DEBUG oslo_vmware.api [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for the task: (returnval){ [ 786.030677] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c9f374-920b-585d-7b90-22d94b41f96c" [ 786.030677] env[62385]: _type = "Task" [ 786.030677] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.057615] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Reconfiguring VM instance instance-0000003a to attach disk [datastore1] 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7/8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 786.058611] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1cbc7289-936d-432a-965a-75306455e112 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.080050] env[62385]: DEBUG oslo_vmware.api [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c9f374-920b-585d-7b90-22d94b41f96c, 'name': SearchDatastore_Task, 'duration_secs': 0.018435} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.080829] env[62385]: DEBUG oslo_concurrency.lockutils [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 786.081096] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 786.081368] env[62385]: DEBUG oslo_concurrency.lockutils [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.081522] env[62385]: DEBUG oslo_concurrency.lockutils [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.081710] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 786.082049] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.082370] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 786.082624] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-15b24019-1221-4336-b089-82967ab83d47 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.084629] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-55514863-dc3e-4688-b1a0-13cde0246202 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.088547] env[62385]: DEBUG oslo_vmware.api [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Waiting for the task: (returnval){ [ 786.088547] env[62385]: value = "task-1205704" [ 786.088547] env[62385]: _type = "Task" [ 786.088547] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.093169] env[62385]: DEBUG oslo_vmware.api [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Waiting for the task: (returnval){ [ 786.093169] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]526c3055-b910-b4e0-5615-2446e9ee67aa" [ 786.093169] env[62385]: _type = "Task" [ 786.093169] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.101584] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3fabae9a-6ac4-4207-8bfe-fd8b2fbb99e2 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.102721] env[62385]: DEBUG oslo_vmware.api [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': task-1205704, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.103015] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 786.103203] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 786.104588] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.559s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.106646] env[62385]: INFO nova.compute.claims [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 786.109811] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb57f12c-dd75-4abf-ae61-5ed9f9c929d3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.116122] env[62385]: DEBUG oslo_vmware.api [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]526c3055-b910-b4e0-5615-2446e9ee67aa, 'name': SearchDatastore_Task, 'duration_secs': 0.013985} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.116839] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 786.117114] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 786.117345] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.121052] env[62385]: DEBUG oslo_concurrency.lockutils [None req-429c844f-60ee-4617-8dd8-4c14c52a8b9a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "0d823fcc-cbba-41f6-a47c-2f57d3e79948" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 100.318s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.121729] env[62385]: DEBUG oslo_vmware.api [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for the task: (returnval){ [ 786.121729] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c35fa2-be1b-d4ba-1fa2-d33f25ae90d7" [ 786.121729] env[62385]: _type = "Task" [ 786.121729] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.132955] env[62385]: DEBUG oslo_vmware.api [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c35fa2-be1b-d4ba-1fa2-d33f25ae90d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.223586] env[62385]: DEBUG nova.network.neutron [-] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.605123] env[62385]: DEBUG oslo_vmware.api [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': task-1205704, 'name': ReconfigVM_Task, 'duration_secs': 0.349404} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.605492] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Reconfigured VM instance instance-0000003a to attach disk [datastore1] 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7/8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 786.606066] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-754b182b-7b54-4420-8c13-900b5b1f163e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.613660] env[62385]: DEBUG oslo_vmware.api [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Waiting for the task: (returnval){ [ 786.613660] env[62385]: value = "task-1205705" [ 786.613660] env[62385]: _type = "Task" [ 786.613660] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.627693] env[62385]: DEBUG oslo_vmware.api [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': task-1205705, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.629192] env[62385]: DEBUG nova.compute.manager [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 786.638750] env[62385]: DEBUG oslo_vmware.api [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c35fa2-be1b-d4ba-1fa2-d33f25ae90d7, 'name': SearchDatastore_Task, 'duration_secs': 0.017929} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.639583] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-393e3297-efb1-469d-8cf6-af17120d79ed {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.645710] env[62385]: DEBUG oslo_vmware.api [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for the task: (returnval){ [ 786.645710] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5231f50b-bf27-d171-40c6-666b4cefd8c6" [ 786.645710] env[62385]: _type = "Task" [ 786.645710] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.657030] env[62385]: DEBUG oslo_vmware.api [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5231f50b-bf27-d171-40c6-666b4cefd8c6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.663170] env[62385]: DEBUG nova.network.neutron [req-9903f7e5-b9a9-4541-8ec2-7ccb64a26e2c req-09ede329-8e9a-4286-b290-3df1ad59d89a service nova] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Updated VIF entry in instance network info cache for port b838271e-07e8-4880-980a-703c1bbd4da5. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 786.663544] env[62385]: DEBUG nova.network.neutron [req-9903f7e5-b9a9-4541-8ec2-7ccb64a26e2c req-09ede329-8e9a-4286-b290-3df1ad59d89a service nova] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Updating instance_info_cache with network_info: [{"id": "b838271e-07e8-4880-980a-703c1bbd4da5", "address": "fa:16:3e:f6:2c:f9", "network": {"id": "1c117704-eb0e-43fc-84fe-5e5059cbb310", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1332686438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "395adff2d8204123b16bf9e8f5c04d05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56834f67-27a8-43dc-bbc6-a74aaa08959b", "external-id": "nsx-vlan-transportzone-949", "segmentation_id": 949, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb838271e-07", "ovs_interfaceid": "b838271e-07e8-4880-980a-703c1bbd4da5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.732180] env[62385]: INFO nova.compute.manager [-] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Took 1.57 seconds to deallocate network for instance. [ 787.131205] env[62385]: DEBUG oslo_vmware.api [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': task-1205705, 'name': Rename_Task, 'duration_secs': 0.204456} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.131205] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 787.131205] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6a1d3efe-c812-46c9-8e48-49c4086c7bf4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.143913] env[62385]: DEBUG oslo_vmware.api [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Waiting for the task: (returnval){ [ 787.143913] env[62385]: value = "task-1205706" [ 787.143913] env[62385]: _type = "Task" [ 787.143913] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.160112] env[62385]: DEBUG oslo_vmware.api [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': task-1205706, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.162836] env[62385]: DEBUG oslo_vmware.api [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5231f50b-bf27-d171-40c6-666b4cefd8c6, 'name': SearchDatastore_Task, 'duration_secs': 0.016479} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.163108] env[62385]: DEBUG oslo_concurrency.lockutils [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.163388] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] b4970e1a-2b29-44e9-b79e-cda3c32dddcd/b4970e1a-2b29-44e9-b79e-cda3c32dddcd.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 787.163676] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.163862] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 787.165689] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bdee5442-3572-402e-b1a4-60dd4df3d9db {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.169662] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.169925] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0a3b63bd-09e9-48d5-b2de-e8345a3b6ac8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.172824] env[62385]: DEBUG oslo_concurrency.lockutils [req-9903f7e5-b9a9-4541-8ec2-7ccb64a26e2c req-09ede329-8e9a-4286-b290-3df1ad59d89a service nova] Releasing lock "refresh_cache-65bb1d14-84b2-4e75-acdc-dc674a035101" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.172824] env[62385]: DEBUG nova.compute.manager [req-9903f7e5-b9a9-4541-8ec2-7ccb64a26e2c req-09ede329-8e9a-4286-b290-3df1ad59d89a service nova] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Received event network-changed-b838271e-07e8-4880-980a-703c1bbd4da5 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 787.172824] env[62385]: DEBUG nova.compute.manager [req-9903f7e5-b9a9-4541-8ec2-7ccb64a26e2c req-09ede329-8e9a-4286-b290-3df1ad59d89a service nova] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Refreshing instance network info cache due to event network-changed-b838271e-07e8-4880-980a-703c1bbd4da5. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 787.173085] env[62385]: DEBUG oslo_concurrency.lockutils [req-9903f7e5-b9a9-4541-8ec2-7ccb64a26e2c req-09ede329-8e9a-4286-b290-3df1ad59d89a service nova] Acquiring lock "refresh_cache-65bb1d14-84b2-4e75-acdc-dc674a035101" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.173164] env[62385]: DEBUG oslo_concurrency.lockutils [req-9903f7e5-b9a9-4541-8ec2-7ccb64a26e2c req-09ede329-8e9a-4286-b290-3df1ad59d89a service nova] Acquired lock "refresh_cache-65bb1d14-84b2-4e75-acdc-dc674a035101" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.173332] env[62385]: DEBUG nova.network.neutron [req-9903f7e5-b9a9-4541-8ec2-7ccb64a26e2c req-09ede329-8e9a-4286-b290-3df1ad59d89a service nova] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Refreshing network info cache for port b838271e-07e8-4880-980a-703c1bbd4da5 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 787.184516] env[62385]: DEBUG oslo_vmware.api [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for the task: (returnval){ [ 787.184516] env[62385]: value = "task-1205707" [ 787.184516] env[62385]: _type = "Task" [ 787.184516] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.186773] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 787.186773] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 787.188086] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5dcf2191-bfa6-4661-8470-a666adf066e0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.194876] env[62385]: DEBUG oslo_vmware.api [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205707, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.203490] env[62385]: DEBUG oslo_vmware.api [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Waiting for the task: (returnval){ [ 787.203490] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52fd5dda-5eb1-0d7c-1ef5-511671f5e230" [ 787.203490] env[62385]: _type = "Task" [ 787.203490] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.219010] env[62385]: DEBUG oslo_vmware.api [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52fd5dda-5eb1-0d7c-1ef5-511671f5e230, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.240440] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ebabc504-fb40-4408-abaf-177b5b541c69 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.376301] env[62385]: DEBUG nova.compute.manager [req-553b5066-6017-4a73-b19c-ed5d27ea922d req-eedd0235-080e-4bff-851d-98667ebd5616 service nova] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Received event network-changed-c14a2370-3801-4b30-ad4f-e86a6b5c133d {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 787.376301] env[62385]: DEBUG nova.compute.manager [req-553b5066-6017-4a73-b19c-ed5d27ea922d req-eedd0235-080e-4bff-851d-98667ebd5616 service nova] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Refreshing instance network info cache due to event network-changed-c14a2370-3801-4b30-ad4f-e86a6b5c133d. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 787.376301] env[62385]: DEBUG oslo_concurrency.lockutils [req-553b5066-6017-4a73-b19c-ed5d27ea922d req-eedd0235-080e-4bff-851d-98667ebd5616 service nova] Acquiring lock "refresh_cache-fd7a77e2-6d76-4d91-bdef-e30333247aa9" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.376301] env[62385]: DEBUG oslo_concurrency.lockutils [req-553b5066-6017-4a73-b19c-ed5d27ea922d req-eedd0235-080e-4bff-851d-98667ebd5616 service nova] Acquired lock "refresh_cache-fd7a77e2-6d76-4d91-bdef-e30333247aa9" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.376301] env[62385]: DEBUG nova.network.neutron [req-553b5066-6017-4a73-b19c-ed5d27ea922d req-eedd0235-080e-4bff-851d-98667ebd5616 service nova] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Refreshing network info cache for port c14a2370-3801-4b30-ad4f-e86a6b5c133d {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 787.645679] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84081801-a964-44fc-8849-ef65bc8a19ff {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.664146] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1002130f-9136-46d8-8393-e1abc8d94aa0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.668640] env[62385]: DEBUG oslo_vmware.api [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': task-1205706, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.718062] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f042580d-58fb-46bc-a816-b6b8b51ae594 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.731468] env[62385]: DEBUG oslo_vmware.api [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205707, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.738905] env[62385]: DEBUG oslo_vmware.api [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52fd5dda-5eb1-0d7c-1ef5-511671f5e230, 'name': SearchDatastore_Task, 'duration_secs': 0.017446} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.740363] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e65519e7-07a3-4726-bbe6-3104356534f0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.747028] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e37ca74e-f955-487f-9c1f-738660b8e0d3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.758145] env[62385]: DEBUG oslo_vmware.api [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Waiting for the task: (returnval){ [ 787.758145] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]525d52d7-f031-c9c1-81ab-876299529f68" [ 787.758145] env[62385]: _type = "Task" [ 787.758145] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.781582] env[62385]: DEBUG nova.compute.provider_tree [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 787.791935] env[62385]: DEBUG oslo_vmware.api [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]525d52d7-f031-c9c1-81ab-876299529f68, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.002313] env[62385]: DEBUG nova.network.neutron [req-9903f7e5-b9a9-4541-8ec2-7ccb64a26e2c req-09ede329-8e9a-4286-b290-3df1ad59d89a service nova] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Updated VIF entry in instance network info cache for port b838271e-07e8-4880-980a-703c1bbd4da5. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 788.002713] env[62385]: DEBUG nova.network.neutron [req-9903f7e5-b9a9-4541-8ec2-7ccb64a26e2c req-09ede329-8e9a-4286-b290-3df1ad59d89a service nova] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Updating instance_info_cache with network_info: [{"id": "b838271e-07e8-4880-980a-703c1bbd4da5", "address": "fa:16:3e:f6:2c:f9", "network": {"id": "1c117704-eb0e-43fc-84fe-5e5059cbb310", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1332686438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "395adff2d8204123b16bf9e8f5c04d05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56834f67-27a8-43dc-bbc6-a74aaa08959b", "external-id": "nsx-vlan-transportzone-949", "segmentation_id": 949, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb838271e-07", "ovs_interfaceid": "b838271e-07e8-4880-980a-703c1bbd4da5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.170090] env[62385]: DEBUG oslo_vmware.api [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': task-1205706, 'name': PowerOnVM_Task, 'duration_secs': 0.693186} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.170418] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 788.171273] env[62385]: DEBUG nova.compute.manager [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 788.171483] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c124db36-d53b-48c1-b8bd-9e1a65c76498 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.220471] env[62385]: DEBUG oslo_vmware.api [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205707, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.720273} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.220759] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] b4970e1a-2b29-44e9-b79e-cda3c32dddcd/b4970e1a-2b29-44e9-b79e-cda3c32dddcd.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 788.220981] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 788.221268] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-25de9f6f-1a6c-4f0d-9697-ac7ff40382a9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.230825] env[62385]: DEBUG oslo_vmware.api [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for the task: (returnval){ [ 788.230825] env[62385]: value = "task-1205708" [ 788.230825] env[62385]: _type = "Task" [ 788.230825] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.243161] env[62385]: DEBUG oslo_vmware.api [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205708, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.266528] env[62385]: DEBUG oslo_vmware.api [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]525d52d7-f031-c9c1-81ab-876299529f68, 'name': SearchDatastore_Task, 'duration_secs': 0.062912} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.266819] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.267107] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] fd7a77e2-6d76-4d91-bdef-e30333247aa9/fd7a77e2-6d76-4d91-bdef-e30333247aa9.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 788.267384] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a91df6df-78ca-4326-b60b-10e1d95466f4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.278444] env[62385]: DEBUG oslo_vmware.api [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Waiting for the task: (returnval){ [ 788.278444] env[62385]: value = "task-1205709" [ 788.278444] env[62385]: _type = "Task" [ 788.278444] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.284984] env[62385]: DEBUG nova.scheduler.client.report [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 788.292758] env[62385]: DEBUG oslo_vmware.api [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Task: {'id': task-1205709, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.480851] env[62385]: DEBUG nova.network.neutron [req-553b5066-6017-4a73-b19c-ed5d27ea922d req-eedd0235-080e-4bff-851d-98667ebd5616 service nova] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Updated VIF entry in instance network info cache for port c14a2370-3801-4b30-ad4f-e86a6b5c133d. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 788.481298] env[62385]: DEBUG nova.network.neutron [req-553b5066-6017-4a73-b19c-ed5d27ea922d req-eedd0235-080e-4bff-851d-98667ebd5616 service nova] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Updating instance_info_cache with network_info: [{"id": "c14a2370-3801-4b30-ad4f-e86a6b5c133d", "address": "fa:16:3e:77:22:db", "network": {"id": "ae13fe4b-9c18-45eb-8619-c58d283abd3f", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-432568731-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83d39da6eb0541fea18f25acc9df3eef", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc14a2370-38", "ovs_interfaceid": "c14a2370-3801-4b30-ad4f-e86a6b5c133d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.506329] env[62385]: DEBUG oslo_concurrency.lockutils [req-9903f7e5-b9a9-4541-8ec2-7ccb64a26e2c req-09ede329-8e9a-4286-b290-3df1ad59d89a service nova] Releasing lock "refresh_cache-65bb1d14-84b2-4e75-acdc-dc674a035101" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.563390] env[62385]: DEBUG nova.compute.manager [req-b040fc2f-2b74-4504-8c2a-ef78234bcc4e req-e2b6578a-8c5f-45d1-82e2-8c0f3bbabfc2 service nova] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Received event network-changed-ad929652-f520-465a-aeea-1db4e8f36fc3 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 788.566024] env[62385]: DEBUG nova.compute.manager [req-b040fc2f-2b74-4504-8c2a-ef78234bcc4e req-e2b6578a-8c5f-45d1-82e2-8c0f3bbabfc2 service nova] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Refreshing instance network info cache due to event network-changed-ad929652-f520-465a-aeea-1db4e8f36fc3. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 788.566024] env[62385]: DEBUG oslo_concurrency.lockutils [req-b040fc2f-2b74-4504-8c2a-ef78234bcc4e req-e2b6578a-8c5f-45d1-82e2-8c0f3bbabfc2 service nova] Acquiring lock "refresh_cache-0d823fcc-cbba-41f6-a47c-2f57d3e79948" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.566024] env[62385]: DEBUG oslo_concurrency.lockutils [req-b040fc2f-2b74-4504-8c2a-ef78234bcc4e req-e2b6578a-8c5f-45d1-82e2-8c0f3bbabfc2 service nova] Acquired lock "refresh_cache-0d823fcc-cbba-41f6-a47c-2f57d3e79948" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.566024] env[62385]: DEBUG nova.network.neutron [req-b040fc2f-2b74-4504-8c2a-ef78234bcc4e req-e2b6578a-8c5f-45d1-82e2-8c0f3bbabfc2 service nova] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Refreshing network info cache for port ad929652-f520-465a-aeea-1db4e8f36fc3 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 788.687585] env[62385]: DEBUG oslo_vmware.rw_handles [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527c1b39-e62c-e7f3-a584-319f831f45e5/disk-0.vmdk. {{(pid=62385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 788.688753] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91dc68be-9cca-40af-adf6-3d97a3f7fdce {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.700019] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.701413] env[62385]: DEBUG oslo_vmware.rw_handles [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527c1b39-e62c-e7f3-a584-319f831f45e5/disk-0.vmdk is in state: ready. {{(pid=62385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 788.702061] env[62385]: ERROR oslo_vmware.rw_handles [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527c1b39-e62c-e7f3-a584-319f831f45e5/disk-0.vmdk due to incomplete transfer. [ 788.702061] env[62385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-155a1315-5ded-4679-88a6-b6edca3fa70f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.711573] env[62385]: DEBUG oslo_vmware.rw_handles [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527c1b39-e62c-e7f3-a584-319f831f45e5/disk-0.vmdk. {{(pid=62385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 788.711573] env[62385]: DEBUG nova.virt.vmwareapi.images [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Uploaded image 682eb0ec-9dc8-46f4-9a39-28fdbc76a00a to the Glance image server {{(pid=62385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 788.714461] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Destroying the VM {{(pid=62385) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 788.714779] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-de6e7706-117d-4f03-9f62-3eb9c906f122 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.722943] env[62385]: DEBUG oslo_vmware.api [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 788.722943] env[62385]: value = "task-1205710" [ 788.722943] env[62385]: _type = "Task" [ 788.722943] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.735578] env[62385]: DEBUG oslo_vmware.api [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205710, 'name': Destroy_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.753827] env[62385]: DEBUG oslo_vmware.api [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205708, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.154205} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.754206] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 788.755523] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-594ae8a0-d295-41e0-906d-cc05625c18cd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.793273] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Reconfiguring VM instance instance-0000003c to attach disk [datastore1] b4970e1a-2b29-44e9-b79e-cda3c32dddcd/b4970e1a-2b29-44e9-b79e-cda3c32dddcd.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 788.798395] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.694s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.799526] env[62385]: DEBUG nova.compute.manager [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 788.804589] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-203b0be7-5ab9-4742-a656-acaa74e16f73 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.819864] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a40d5bae-eeab-45c3-9849-2e493ad05ea0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.292s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.820147] env[62385]: DEBUG nova.objects.instance [None req-a40d5bae-eeab-45c3-9849-2e493ad05ea0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lazy-loading 'resources' on Instance uuid caca5304-0ad7-427c-812e-de925de63f2c {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 788.828602] env[62385]: DEBUG oslo_vmware.api [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Task: {'id': task-1205709, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.830294] env[62385]: DEBUG oslo_vmware.api [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for the task: (returnval){ [ 788.830294] env[62385]: value = "task-1205711" [ 788.830294] env[62385]: _type = "Task" [ 788.830294] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.841101] env[62385]: DEBUG oslo_vmware.api [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205711, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.983860] env[62385]: DEBUG oslo_concurrency.lockutils [req-553b5066-6017-4a73-b19c-ed5d27ea922d req-eedd0235-080e-4bff-851d-98667ebd5616 service nova] Releasing lock "refresh_cache-fd7a77e2-6d76-4d91-bdef-e30333247aa9" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.984254] env[62385]: DEBUG nova.compute.manager [req-553b5066-6017-4a73-b19c-ed5d27ea922d req-eedd0235-080e-4bff-851d-98667ebd5616 service nova] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Received event network-vif-deleted-fdae3362-d2d0-4a5e-80ff-c92e442ef9f4 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 789.236717] env[62385]: DEBUG oslo_vmware.api [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205710, 'name': Destroy_Task} progress is 33%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.293020] env[62385]: DEBUG oslo_vmware.api [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Task: {'id': task-1205709, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.918616} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.293020] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] fd7a77e2-6d76-4d91-bdef-e30333247aa9/fd7a77e2-6d76-4d91-bdef-e30333247aa9.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 789.293020] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 789.293248] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-20228f01-951c-44b7-b2dc-8b21408e6468 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.302112] env[62385]: DEBUG oslo_vmware.api [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Waiting for the task: (returnval){ [ 789.302112] env[62385]: value = "task-1205712" [ 789.302112] env[62385]: _type = "Task" [ 789.302112] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.309692] env[62385]: DEBUG nova.compute.utils [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 789.312364] env[62385]: DEBUG nova.compute.manager [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 789.312608] env[62385]: DEBUG nova.network.neutron [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 789.325592] env[62385]: DEBUG oslo_vmware.api [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Task: {'id': task-1205712, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.343683] env[62385]: DEBUG oslo_vmware.api [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205711, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.370246] env[62385]: DEBUG nova.policy [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8988632e3af942a884b9c811e4ca539c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6e99f1e4cfd84bf083c8261b2ffcfa2f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 789.495066] env[62385]: DEBUG nova.network.neutron [req-b040fc2f-2b74-4504-8c2a-ef78234bcc4e req-e2b6578a-8c5f-45d1-82e2-8c0f3bbabfc2 service nova] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Updated VIF entry in instance network info cache for port ad929652-f520-465a-aeea-1db4e8f36fc3. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 789.495293] env[62385]: DEBUG nova.network.neutron [req-b040fc2f-2b74-4504-8c2a-ef78234bcc4e req-e2b6578a-8c5f-45d1-82e2-8c0f3bbabfc2 service nova] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Updating instance_info_cache with network_info: [{"id": "ad929652-f520-465a-aeea-1db4e8f36fc3", "address": "fa:16:3e:b6:00:7c", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.130", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad929652-f5", "ovs_interfaceid": "ad929652-f520-465a-aeea-1db4e8f36fc3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.737499] env[62385]: DEBUG oslo_vmware.api [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205710, 'name': Destroy_Task, 'duration_secs': 0.923051} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.738971] env[62385]: DEBUG nova.network.neutron [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Successfully created port: 2c31d161-814f-4c23-9882-0aa436bb9e03 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 789.739962] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Destroyed the VM [ 789.740152] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Deleting Snapshot of the VM instance {{(pid=62385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 789.741305] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-fc52e943-e64d-41a5-9619-883e8b37b7ce {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.745023] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2696deda-9caf-4da5-92da-f82fcb0b98c2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.753814] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a73fbba9-a3fd-4598-99a0-3ff3af9b867e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.757317] env[62385]: DEBUG oslo_vmware.api [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 789.757317] env[62385]: value = "task-1205713" [ 789.757317] env[62385]: _type = "Task" [ 789.757317] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.787892] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3a2b0cd-2f69-424c-b05e-29b46a11cae0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.794032] env[62385]: DEBUG oslo_vmware.api [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205713, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.799854] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faec9862-c2f2-4e05-b7ee-7f258d57e7f3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.815917] env[62385]: DEBUG oslo_vmware.api [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Task: {'id': task-1205712, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.16547} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.823568] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 789.824168] env[62385]: DEBUG nova.compute.manager [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 789.827039] env[62385]: DEBUG nova.compute.provider_tree [None req-a40d5bae-eeab-45c3-9849-2e493ad05ea0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 789.828927] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f50072db-2de4-4c38-9204-fe5c1487b57b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.852314] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] fd7a77e2-6d76-4d91-bdef-e30333247aa9/fd7a77e2-6d76-4d91-bdef-e30333247aa9.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 789.853679] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0cc64df0-a67e-4daa-b78c-fade12cd0655 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.872224] env[62385]: DEBUG oslo_vmware.api [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205711, 'name': ReconfigVM_Task, 'duration_secs': 0.537945} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.873141] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Reconfigured VM instance instance-0000003c to attach disk [datastore1] b4970e1a-2b29-44e9-b79e-cda3c32dddcd/b4970e1a-2b29-44e9-b79e-cda3c32dddcd.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 789.873595] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e1a225f6-8916-4446-8721-a0fa1ebcbef9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.878933] env[62385]: DEBUG oslo_vmware.api [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Waiting for the task: (returnval){ [ 789.878933] env[62385]: value = "task-1205714" [ 789.878933] env[62385]: _type = "Task" [ 789.878933] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.883428] env[62385]: DEBUG oslo_vmware.api [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for the task: (returnval){ [ 789.883428] env[62385]: value = "task-1205715" [ 789.883428] env[62385]: _type = "Task" [ 789.883428] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.891406] env[62385]: DEBUG oslo_vmware.api [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Task: {'id': task-1205714, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.894610] env[62385]: DEBUG oslo_vmware.api [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205715, 'name': Rename_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.998611] env[62385]: DEBUG oslo_concurrency.lockutils [req-b040fc2f-2b74-4504-8c2a-ef78234bcc4e req-e2b6578a-8c5f-45d1-82e2-8c0f3bbabfc2 service nova] Releasing lock "refresh_cache-0d823fcc-cbba-41f6-a47c-2f57d3e79948" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.998873] env[62385]: DEBUG nova.compute.manager [req-b040fc2f-2b74-4504-8c2a-ef78234bcc4e req-e2b6578a-8c5f-45d1-82e2-8c0f3bbabfc2 service nova] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Received event network-changed-b838271e-07e8-4880-980a-703c1bbd4da5 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 789.999065] env[62385]: DEBUG nova.compute.manager [req-b040fc2f-2b74-4504-8c2a-ef78234bcc4e req-e2b6578a-8c5f-45d1-82e2-8c0f3bbabfc2 service nova] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Refreshing instance network info cache due to event network-changed-b838271e-07e8-4880-980a-703c1bbd4da5. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 789.999324] env[62385]: DEBUG oslo_concurrency.lockutils [req-b040fc2f-2b74-4504-8c2a-ef78234bcc4e req-e2b6578a-8c5f-45d1-82e2-8c0f3bbabfc2 service nova] Acquiring lock "refresh_cache-65bb1d14-84b2-4e75-acdc-dc674a035101" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.999481] env[62385]: DEBUG oslo_concurrency.lockutils [req-b040fc2f-2b74-4504-8c2a-ef78234bcc4e req-e2b6578a-8c5f-45d1-82e2-8c0f3bbabfc2 service nova] Acquired lock "refresh_cache-65bb1d14-84b2-4e75-acdc-dc674a035101" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.999651] env[62385]: DEBUG nova.network.neutron [req-b040fc2f-2b74-4504-8c2a-ef78234bcc4e req-e2b6578a-8c5f-45d1-82e2-8c0f3bbabfc2 service nova] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Refreshing network info cache for port b838271e-07e8-4880-980a-703c1bbd4da5 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 790.268782] env[62385]: DEBUG oslo_vmware.api [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205713, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.336623] env[62385]: DEBUG nova.scheduler.client.report [None req-a40d5bae-eeab-45c3-9849-2e493ad05ea0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 790.396257] env[62385]: DEBUG oslo_vmware.api [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Task: {'id': task-1205714, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.403717] env[62385]: DEBUG oslo_vmware.api [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205715, 'name': Rename_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.526427] env[62385]: DEBUG nova.compute.manager [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 790.527434] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a18a0b78-66a0-4a17-bf99-b94e982cac56 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.768803] env[62385]: DEBUG oslo_vmware.api [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205713, 'name': RemoveSnapshot_Task, 'duration_secs': 0.800726} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.771795] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Deleted Snapshot of the VM instance {{(pid=62385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 790.772095] env[62385]: DEBUG nova.compute.manager [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 790.777826] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1734a6fe-7853-402e-8317-38d4e2885ae4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.813019] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Acquiring lock "7f2fe830-e16d-4684-91a4-aab219468e77" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.813284] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Lock "7f2fe830-e16d-4684-91a4-aab219468e77" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.835570] env[62385]: DEBUG nova.compute.manager [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 790.846260] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a40d5bae-eeab-45c3-9849-2e493ad05ea0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.026s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.850622] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c3bb0442-402a-44c3-a5ec-63b1cb851d77 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.418s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.850622] env[62385]: DEBUG nova.objects.instance [None req-c3bb0442-402a-44c3-a5ec-63b1cb851d77 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lazy-loading 'resources' on Instance uuid 42ecf594-dff7-4af3-ac56-5bbc45b7a192 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 790.858678] env[62385]: DEBUG nova.virt.hardware [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 790.858924] env[62385]: DEBUG nova.virt.hardware [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 790.859102] env[62385]: DEBUG nova.virt.hardware [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 790.859379] env[62385]: DEBUG nova.virt.hardware [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 790.859557] env[62385]: DEBUG nova.virt.hardware [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 790.859727] env[62385]: DEBUG nova.virt.hardware [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 790.859947] env[62385]: DEBUG nova.virt.hardware [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 790.860302] env[62385]: DEBUG nova.virt.hardware [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 790.860572] env[62385]: DEBUG nova.virt.hardware [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 790.860696] env[62385]: DEBUG nova.virt.hardware [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 790.860874] env[62385]: DEBUG nova.virt.hardware [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 790.862409] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-853b92e3-e6a3-4b12-960c-e8256a1a959e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.872696] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee8efa9b-0961-4388-9904-08551288f92a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.877395] env[62385]: INFO nova.scheduler.client.report [None req-a40d5bae-eeab-45c3-9849-2e493ad05ea0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Deleted allocations for instance caca5304-0ad7-427c-812e-de925de63f2c [ 790.914291] env[62385]: DEBUG oslo_vmware.api [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205715, 'name': Rename_Task, 'duration_secs': 0.739058} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.919607] env[62385]: DEBUG nova.network.neutron [req-b040fc2f-2b74-4504-8c2a-ef78234bcc4e req-e2b6578a-8c5f-45d1-82e2-8c0f3bbabfc2 service nova] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Updated VIF entry in instance network info cache for port b838271e-07e8-4880-980a-703c1bbd4da5. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 790.920024] env[62385]: DEBUG nova.network.neutron [req-b040fc2f-2b74-4504-8c2a-ef78234bcc4e req-e2b6578a-8c5f-45d1-82e2-8c0f3bbabfc2 service nova] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Updating instance_info_cache with network_info: [{"id": "b838271e-07e8-4880-980a-703c1bbd4da5", "address": "fa:16:3e:f6:2c:f9", "network": {"id": "1c117704-eb0e-43fc-84fe-5e5059cbb310", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1332686438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "395adff2d8204123b16bf9e8f5c04d05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56834f67-27a8-43dc-bbc6-a74aaa08959b", "external-id": "nsx-vlan-transportzone-949", "segmentation_id": 949, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb838271e-07", "ovs_interfaceid": "b838271e-07e8-4880-980a-703c1bbd4da5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.921417] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 790.921784] env[62385]: DEBUG oslo_vmware.api [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Task: {'id': task-1205714, 'name': ReconfigVM_Task, 'duration_secs': 0.899606} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.922633] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e410709c-8e51-46b2-9c54-48824f58b016 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.924722] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Reconfigured VM instance instance-0000003d to attach disk [datastore1] fd7a77e2-6d76-4d91-bdef-e30333247aa9/fd7a77e2-6d76-4d91-bdef-e30333247aa9.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 790.925834] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fc1e534b-4324-46b9-b970-11ba0392b6f1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.934510] env[62385]: DEBUG oslo_vmware.api [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Waiting for the task: (returnval){ [ 790.934510] env[62385]: value = "task-1205717" [ 790.934510] env[62385]: _type = "Task" [ 790.934510] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.936454] env[62385]: DEBUG oslo_vmware.api [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for the task: (returnval){ [ 790.936454] env[62385]: value = "task-1205716" [ 790.936454] env[62385]: _type = "Task" [ 790.936454] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.949135] env[62385]: DEBUG oslo_vmware.api [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Task: {'id': task-1205717, 'name': Rename_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.952170] env[62385]: DEBUG oslo_vmware.api [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205716, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.031575] env[62385]: DEBUG nova.compute.manager [req-5f54552d-69fa-40b5-a0cf-0f40af027397 req-5987e44e-4d44-47fd-a241-c47eaddf9343 service nova] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Received event network-changed-b838271e-07e8-4880-980a-703c1bbd4da5 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 791.031774] env[62385]: DEBUG nova.compute.manager [req-5f54552d-69fa-40b5-a0cf-0f40af027397 req-5987e44e-4d44-47fd-a241-c47eaddf9343 service nova] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Refreshing instance network info cache due to event network-changed-b838271e-07e8-4880-980a-703c1bbd4da5. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 791.031974] env[62385]: DEBUG oslo_concurrency.lockutils [req-5f54552d-69fa-40b5-a0cf-0f40af027397 req-5987e44e-4d44-47fd-a241-c47eaddf9343 service nova] Acquiring lock "refresh_cache-65bb1d14-84b2-4e75-acdc-dc674a035101" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.038667] env[62385]: INFO nova.compute.manager [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] instance snapshotting [ 791.041464] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2ebda03-35b5-40e0-8495-3a9ef227fcc1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.063111] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60dafd5f-dd3e-4596-a90a-724ae2207c02 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.290648] env[62385]: INFO nova.compute.manager [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Shelve offloading [ 791.291157] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 791.291592] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-decc64c7-4734-40b7-804e-adba6fd58dde {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.302137] env[62385]: DEBUG oslo_vmware.api [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 791.302137] env[62385]: value = "task-1205718" [ 791.302137] env[62385]: _type = "Task" [ 791.302137] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.314299] env[62385]: DEBUG nova.network.neutron [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Successfully updated port: 2c31d161-814f-4c23-9882-0aa436bb9e03 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 791.314875] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] VM already powered off {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 791.315077] env[62385]: DEBUG nova.compute.manager [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 791.316080] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6986075-d316-4d6d-aed8-16ad4688e658 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.322889] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "refresh_cache-f8a98710-9d02-4aa3-b6a5-34b7060ad62d" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.323096] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquired lock "refresh_cache-f8a98710-9d02-4aa3-b6a5-34b7060ad62d" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.323273] env[62385]: DEBUG nova.network.neutron [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 791.388873] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a40d5bae-eeab-45c3-9849-2e493ad05ea0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "caca5304-0ad7-427c-812e-de925de63f2c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.638s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.394335] env[62385]: DEBUG nova.compute.manager [req-1d2e2266-8b53-453e-8f5f-765452e0cbd7 req-21bf9ea2-8713-4147-9bcd-453b24945590 service nova] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Received event network-vif-plugged-2c31d161-814f-4c23-9882-0aa436bb9e03 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 791.394335] env[62385]: DEBUG oslo_concurrency.lockutils [req-1d2e2266-8b53-453e-8f5f-765452e0cbd7 req-21bf9ea2-8713-4147-9bcd-453b24945590 service nova] Acquiring lock "d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.394335] env[62385]: DEBUG oslo_concurrency.lockutils [req-1d2e2266-8b53-453e-8f5f-765452e0cbd7 req-21bf9ea2-8713-4147-9bcd-453b24945590 service nova] Lock "d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.394335] env[62385]: DEBUG oslo_concurrency.lockutils [req-1d2e2266-8b53-453e-8f5f-765452e0cbd7 req-21bf9ea2-8713-4147-9bcd-453b24945590 service nova] Lock "d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.394335] env[62385]: DEBUG nova.compute.manager [req-1d2e2266-8b53-453e-8f5f-765452e0cbd7 req-21bf9ea2-8713-4147-9bcd-453b24945590 service nova] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] No waiting events found dispatching network-vif-plugged-2c31d161-814f-4c23-9882-0aa436bb9e03 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 791.394730] env[62385]: WARNING nova.compute.manager [req-1d2e2266-8b53-453e-8f5f-765452e0cbd7 req-21bf9ea2-8713-4147-9bcd-453b24945590 service nova] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Received unexpected event network-vif-plugged-2c31d161-814f-4c23-9882-0aa436bb9e03 for instance with vm_state building and task_state spawning. [ 791.405835] env[62385]: INFO nova.compute.manager [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Rebuilding instance [ 791.424021] env[62385]: DEBUG oslo_concurrency.lockutils [req-b040fc2f-2b74-4504-8c2a-ef78234bcc4e req-e2b6578a-8c5f-45d1-82e2-8c0f3bbabfc2 service nova] Releasing lock "refresh_cache-65bb1d14-84b2-4e75-acdc-dc674a035101" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.424139] env[62385]: DEBUG oslo_concurrency.lockutils [req-5f54552d-69fa-40b5-a0cf-0f40af027397 req-5987e44e-4d44-47fd-a241-c47eaddf9343 service nova] Acquired lock "refresh_cache-65bb1d14-84b2-4e75-acdc-dc674a035101" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.424348] env[62385]: DEBUG nova.network.neutron [req-5f54552d-69fa-40b5-a0cf-0f40af027397 req-5987e44e-4d44-47fd-a241-c47eaddf9343 service nova] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Refreshing network info cache for port b838271e-07e8-4880-980a-703c1bbd4da5 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 791.458267] env[62385]: DEBUG oslo_vmware.api [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Task: {'id': task-1205717, 'name': Rename_Task, 'duration_secs': 0.191762} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.458638] env[62385]: DEBUG oslo_vmware.api [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205716, 'name': PowerOnVM_Task, 'duration_secs': 0.446983} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.461380] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 791.461380] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 791.461380] env[62385]: INFO nova.compute.manager [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Took 11.23 seconds to spawn the instance on the hypervisor. [ 791.461583] env[62385]: DEBUG nova.compute.manager [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 791.462238] env[62385]: DEBUG nova.compute.manager [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 791.462471] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d346bfa1-d4f7-47a9-b73f-9df0ae4d251c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.464547] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7500f493-ad57-463f-b62a-6b3af05a331b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.467730] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86dc4f45-bba2-4d45-b75b-81e69ca301ca {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.484885] env[62385]: DEBUG oslo_vmware.api [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Waiting for the task: (returnval){ [ 791.484885] env[62385]: value = "task-1205719" [ 791.484885] env[62385]: _type = "Task" [ 791.484885] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.497518] env[62385]: DEBUG oslo_vmware.api [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Task: {'id': task-1205719, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.576094] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Creating Snapshot of the VM instance {{(pid=62385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 791.576871] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-8f620f25-8920-42b4-aa50-036b80f87f7b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.586158] env[62385]: DEBUG oslo_vmware.api [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Waiting for the task: (returnval){ [ 791.586158] env[62385]: value = "task-1205720" [ 791.586158] env[62385]: _type = "Task" [ 791.586158] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.599190] env[62385]: DEBUG oslo_vmware.api [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205720, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.775717] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3674d5b9-8921-4588-9200-1f86474a253c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.784355] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23f7a3c0-1bbf-4b72-a9c0-54218d99130b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.816341] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Acquiring lock "refresh_cache-d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.816466] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Acquired lock "refresh_cache-d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.816601] env[62385]: DEBUG nova.network.neutron [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 791.818540] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4486a6de-fcfe-4538-9b7e-b6869cec1a27 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.829816] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd9a8f11-8120-4732-a75b-0c54f34502db {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.853159] env[62385]: DEBUG nova.compute.provider_tree [None req-c3bb0442-402a-44c3-a5ec-63b1cb851d77 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 791.991614] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 791.993268] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9dce91f8-cc83-4a39-aca2-96d8d72dc1c0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.995323] env[62385]: INFO nova.compute.manager [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Took 43.28 seconds to build instance. [ 792.006357] env[62385]: DEBUG oslo_vmware.api [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Task: {'id': task-1205719, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.008277] env[62385]: DEBUG oslo_vmware.api [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Waiting for the task: (returnval){ [ 792.008277] env[62385]: value = "task-1205721" [ 792.008277] env[62385]: _type = "Task" [ 792.008277] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.021834] env[62385]: DEBUG oslo_vmware.api [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Task: {'id': task-1205721, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.083077] env[62385]: DEBUG nova.network.neutron [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Updating instance_info_cache with network_info: [{"id": "bfb1f5cb-27f6-4422-8a40-f8f2d8a6520b", "address": "fa:16:3e:0c:f4:9c", "network": {"id": "ff21dd88-fa02-4bd6-8009-980ec6e72758", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1065789169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aea48968d4464ac0ab01c998c059107d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3ff3baee-99ce-4b51-ae98-efc6163aaab3", "external-id": "nsx-vlan-transportzone-574", "segmentation_id": 574, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbfb1f5cb-27", "ovs_interfaceid": "bfb1f5cb-27f6-4422-8a40-f8f2d8a6520b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.097568] env[62385]: DEBUG oslo_vmware.api [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205720, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.236514] env[62385]: DEBUG oslo_concurrency.lockutils [None req-650314b1-bfcb-451d-9c85-40f84a6535ea tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Acquiring lock "65bb1d14-84b2-4e75-acdc-dc674a035101" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.236800] env[62385]: DEBUG oslo_concurrency.lockutils [None req-650314b1-bfcb-451d-9c85-40f84a6535ea tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Lock "65bb1d14-84b2-4e75-acdc-dc674a035101" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.237025] env[62385]: DEBUG oslo_concurrency.lockutils [None req-650314b1-bfcb-451d-9c85-40f84a6535ea tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Acquiring lock "65bb1d14-84b2-4e75-acdc-dc674a035101-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.237222] env[62385]: DEBUG oslo_concurrency.lockutils [None req-650314b1-bfcb-451d-9c85-40f84a6535ea tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Lock "65bb1d14-84b2-4e75-acdc-dc674a035101-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.237434] env[62385]: DEBUG oslo_concurrency.lockutils [None req-650314b1-bfcb-451d-9c85-40f84a6535ea tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Lock "65bb1d14-84b2-4e75-acdc-dc674a035101-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.240261] env[62385]: INFO nova.compute.manager [None req-650314b1-bfcb-451d-9c85-40f84a6535ea tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Terminating instance [ 792.242126] env[62385]: DEBUG nova.compute.manager [None req-650314b1-bfcb-451d-9c85-40f84a6535ea tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 792.242267] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-650314b1-bfcb-451d-9c85-40f84a6535ea tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 792.243119] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-853440d6-0d3d-47d9-8e35-641ccebb577f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.251167] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-650314b1-bfcb-451d-9c85-40f84a6535ea tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 792.251409] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3785c22e-75f3-4b46-a741-f8179b129288 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.258649] env[62385]: DEBUG oslo_vmware.api [None req-650314b1-bfcb-451d-9c85-40f84a6535ea tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Waiting for the task: (returnval){ [ 792.258649] env[62385]: value = "task-1205722" [ 792.258649] env[62385]: _type = "Task" [ 792.258649] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.266916] env[62385]: DEBUG oslo_vmware.api [None req-650314b1-bfcb-451d-9c85-40f84a6535ea tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': task-1205722, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.271864] env[62385]: DEBUG nova.network.neutron [req-5f54552d-69fa-40b5-a0cf-0f40af027397 req-5987e44e-4d44-47fd-a241-c47eaddf9343 service nova] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Updated VIF entry in instance network info cache for port b838271e-07e8-4880-980a-703c1bbd4da5. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 792.272295] env[62385]: DEBUG nova.network.neutron [req-5f54552d-69fa-40b5-a0cf-0f40af027397 req-5987e44e-4d44-47fd-a241-c47eaddf9343 service nova] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Updating instance_info_cache with network_info: [{"id": "b838271e-07e8-4880-980a-703c1bbd4da5", "address": "fa:16:3e:f6:2c:f9", "network": {"id": "1c117704-eb0e-43fc-84fe-5e5059cbb310", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1332686438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "395adff2d8204123b16bf9e8f5c04d05", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56834f67-27a8-43dc-bbc6-a74aaa08959b", "external-id": "nsx-vlan-transportzone-949", "segmentation_id": 949, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb838271e-07", "ovs_interfaceid": "b838271e-07e8-4880-980a-703c1bbd4da5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.355803] env[62385]: DEBUG nova.scheduler.client.report [None req-c3bb0442-402a-44c3-a5ec-63b1cb851d77 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 792.369693] env[62385]: DEBUG nova.network.neutron [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 792.378467] env[62385]: INFO nova.compute.manager [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Rescuing [ 792.378755] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Acquiring lock "refresh_cache-b4970e1a-2b29-44e9-b79e-cda3c32dddcd" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.378948] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Acquired lock "refresh_cache-b4970e1a-2b29-44e9-b79e-cda3c32dddcd" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.379278] env[62385]: DEBUG nova.network.neutron [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 792.500116] env[62385]: DEBUG oslo_concurrency.lockutils [None req-094a12a8-c38f-4040-a379-2813faf70402 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Lock "b4970e1a-2b29-44e9-b79e-cda3c32dddcd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 97.666s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.505409] env[62385]: DEBUG oslo_vmware.api [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Task: {'id': task-1205719, 'name': PowerOnVM_Task, 'duration_secs': 0.656278} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.505409] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 792.505409] env[62385]: INFO nova.compute.manager [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Took 9.70 seconds to spawn the instance on the hypervisor. [ 792.505409] env[62385]: DEBUG nova.compute.manager [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 792.505975] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-281a16a5-35e8-474c-bcf3-e3c23cdf15bf {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.526804] env[62385]: DEBUG oslo_vmware.api [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Task: {'id': task-1205721, 'name': PowerOffVM_Task, 'duration_secs': 0.131084} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.527483] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 792.527807] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 792.528952] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e59dbe3b-d838-4235-8072-9051922e9ae5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.540018] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 792.540309] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-99578980-9e3a-4fbd-935d-3446231c631e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.569286] env[62385]: DEBUG nova.network.neutron [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Updating instance_info_cache with network_info: [{"id": "2c31d161-814f-4c23-9882-0aa436bb9e03", "address": "fa:16:3e:d7:43:42", "network": {"id": "aca904e1-fbd6-4def-bbec-bd70c0fc8af9", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-249813742-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e99f1e4cfd84bf083c8261b2ffcfa2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f7abe5d-b8fe-4983-bd50-e7469f1fe7f3", "external-id": "nsx-vlan-transportzone-263", "segmentation_id": 263, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c31d161-81", "ovs_interfaceid": "2c31d161-814f-4c23-9882-0aa436bb9e03", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.574307] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 792.574586] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 792.574798] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Deleting the datastore file [datastore1] 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 792.575101] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fd5ecb29-75e7-45c2-b0ee-ed2389273511 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.584675] env[62385]: DEBUG oslo_vmware.api [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Waiting for the task: (returnval){ [ 792.584675] env[62385]: value = "task-1205724" [ 792.584675] env[62385]: _type = "Task" [ 792.584675] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.587846] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Releasing lock "refresh_cache-f8a98710-9d02-4aa3-b6a5-34b7060ad62d" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.598276] env[62385]: DEBUG oslo_vmware.api [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Task: {'id': task-1205724, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.601509] env[62385]: DEBUG oslo_vmware.api [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205720, 'name': CreateSnapshot_Task, 'duration_secs': 0.92731} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.602131] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Created Snapshot of the VM instance {{(pid=62385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 792.602564] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-055a4ed2-1aab-4c37-858b-2d66453358fd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.769985] env[62385]: DEBUG oslo_vmware.api [None req-650314b1-bfcb-451d-9c85-40f84a6535ea tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': task-1205722, 'name': PowerOffVM_Task, 'duration_secs': 0.321287} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.770300] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-650314b1-bfcb-451d-9c85-40f84a6535ea tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 792.770477] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-650314b1-bfcb-451d-9c85-40f84a6535ea tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 792.770736] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a6c0d7b7-0f31-4ec5-8ac3-9d2d90e039ff {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.775283] env[62385]: DEBUG oslo_concurrency.lockutils [req-5f54552d-69fa-40b5-a0cf-0f40af027397 req-5987e44e-4d44-47fd-a241-c47eaddf9343 service nova] Releasing lock "refresh_cache-65bb1d14-84b2-4e75-acdc-dc674a035101" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.846432] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-650314b1-bfcb-451d-9c85-40f84a6535ea tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 792.846746] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-650314b1-bfcb-451d-9c85-40f84a6535ea tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Deleting contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 792.846858] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-650314b1-bfcb-451d-9c85-40f84a6535ea tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Deleting the datastore file [datastore2] 65bb1d14-84b2-4e75-acdc-dc674a035101 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 792.847153] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a1637a2b-d11a-42ef-8052-d277cf781b83 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.854289] env[62385]: DEBUG oslo_vmware.api [None req-650314b1-bfcb-451d-9c85-40f84a6535ea tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Waiting for the task: (returnval){ [ 792.854289] env[62385]: value = "task-1205726" [ 792.854289] env[62385]: _type = "Task" [ 792.854289] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.863581] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c3bb0442-402a-44c3-a5ec-63b1cb851d77 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.015s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.865527] env[62385]: DEBUG oslo_vmware.api [None req-650314b1-bfcb-451d-9c85-40f84a6535ea tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': task-1205726, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.865983] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0e43bfd4-0da1-49c6-a0cb-aa2b4014286e tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.941s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.866237] env[62385]: DEBUG nova.objects.instance [None req-0e43bfd4-0da1-49c6-a0cb-aa2b4014286e tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Lazy-loading 'resources' on Instance uuid 92057af7-28a3-4643-9cda-d3d868d0cfc9 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 792.893216] env[62385]: INFO nova.scheduler.client.report [None req-c3bb0442-402a-44c3-a5ec-63b1cb851d77 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Deleted allocations for instance 42ecf594-dff7-4af3-ac56-5bbc45b7a192 [ 792.977527] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 792.978499] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1119a7ef-eff9-4f1b-ae00-fba3e28f460b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.987429] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 792.987683] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-641f3406-cee8-46b2-831a-60bfcbf1751a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.006179] env[62385]: DEBUG nova.compute.manager [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 793.035808] env[62385]: INFO nova.compute.manager [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Took 43.63 seconds to build instance. [ 793.062251] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 793.062485] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Deleting contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 793.062671] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Deleting the datastore file [datastore2] f8a98710-9d02-4aa3-b6a5-34b7060ad62d {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 793.062939] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-62eca94b-95a8-4894-83a3-697e51f0064c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.069837] env[62385]: DEBUG nova.compute.manager [req-e8d29cdd-6b86-4836-b2b7-409cc1572187 req-872a7895-049b-4fef-a21d-5149d0f7ee8a service nova] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Received event network-vif-unplugged-bfb1f5cb-27f6-4422-8a40-f8f2d8a6520b {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 793.070065] env[62385]: DEBUG oslo_concurrency.lockutils [req-e8d29cdd-6b86-4836-b2b7-409cc1572187 req-872a7895-049b-4fef-a21d-5149d0f7ee8a service nova] Acquiring lock "f8a98710-9d02-4aa3-b6a5-34b7060ad62d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.070280] env[62385]: DEBUG oslo_concurrency.lockutils [req-e8d29cdd-6b86-4836-b2b7-409cc1572187 req-872a7895-049b-4fef-a21d-5149d0f7ee8a service nova] Lock "f8a98710-9d02-4aa3-b6a5-34b7060ad62d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.070903] env[62385]: DEBUG oslo_concurrency.lockutils [req-e8d29cdd-6b86-4836-b2b7-409cc1572187 req-872a7895-049b-4fef-a21d-5149d0f7ee8a service nova] Lock "f8a98710-9d02-4aa3-b6a5-34b7060ad62d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.070903] env[62385]: DEBUG nova.compute.manager [req-e8d29cdd-6b86-4836-b2b7-409cc1572187 req-872a7895-049b-4fef-a21d-5149d0f7ee8a service nova] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] No waiting events found dispatching network-vif-unplugged-bfb1f5cb-27f6-4422-8a40-f8f2d8a6520b {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 793.070903] env[62385]: WARNING nova.compute.manager [req-e8d29cdd-6b86-4836-b2b7-409cc1572187 req-872a7895-049b-4fef-a21d-5149d0f7ee8a service nova] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Received unexpected event network-vif-unplugged-bfb1f5cb-27f6-4422-8a40-f8f2d8a6520b for instance with vm_state shelved and task_state shelving_offloading. [ 793.071106] env[62385]: DEBUG nova.compute.manager [req-e8d29cdd-6b86-4836-b2b7-409cc1572187 req-872a7895-049b-4fef-a21d-5149d0f7ee8a service nova] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Received event network-changed-bfb1f5cb-27f6-4422-8a40-f8f2d8a6520b {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 793.071250] env[62385]: DEBUG nova.compute.manager [req-e8d29cdd-6b86-4836-b2b7-409cc1572187 req-872a7895-049b-4fef-a21d-5149d0f7ee8a service nova] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Refreshing instance network info cache due to event network-changed-bfb1f5cb-27f6-4422-8a40-f8f2d8a6520b. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 793.071585] env[62385]: DEBUG oslo_concurrency.lockutils [req-e8d29cdd-6b86-4836-b2b7-409cc1572187 req-872a7895-049b-4fef-a21d-5149d0f7ee8a service nova] Acquiring lock "refresh_cache-f8a98710-9d02-4aa3-b6a5-34b7060ad62d" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.071918] env[62385]: DEBUG oslo_concurrency.lockutils [req-e8d29cdd-6b86-4836-b2b7-409cc1572187 req-872a7895-049b-4fef-a21d-5149d0f7ee8a service nova] Acquired lock "refresh_cache-f8a98710-9d02-4aa3-b6a5-34b7060ad62d" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.072255] env[62385]: DEBUG nova.network.neutron [req-e8d29cdd-6b86-4836-b2b7-409cc1572187 req-872a7895-049b-4fef-a21d-5149d0f7ee8a service nova] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Refreshing network info cache for port bfb1f5cb-27f6-4422-8a40-f8f2d8a6520b {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 793.075324] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Releasing lock "refresh_cache-d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.075741] env[62385]: DEBUG nova.compute.manager [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Instance network_info: |[{"id": "2c31d161-814f-4c23-9882-0aa436bb9e03", "address": "fa:16:3e:d7:43:42", "network": {"id": "aca904e1-fbd6-4def-bbec-bd70c0fc8af9", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-249813742-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e99f1e4cfd84bf083c8261b2ffcfa2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f7abe5d-b8fe-4983-bd50-e7469f1fe7f3", "external-id": "nsx-vlan-transportzone-263", "segmentation_id": 263, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c31d161-81", "ovs_interfaceid": "2c31d161-814f-4c23-9882-0aa436bb9e03", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 793.076961] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d7:43:42', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2f7abe5d-b8fe-4983-bd50-e7469f1fe7f3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2c31d161-814f-4c23-9882-0aa436bb9e03', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 793.090061] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Creating folder: Project (6e99f1e4cfd84bf083c8261b2ffcfa2f). Parent ref: group-v261107. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 793.092435] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3f2992a9-73f1-4ead-9761-c0829f77e0b6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.099321] env[62385]: DEBUG oslo_vmware.api [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 793.099321] env[62385]: value = "task-1205728" [ 793.099321] env[62385]: _type = "Task" [ 793.099321] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.112804] env[62385]: DEBUG oslo_vmware.api [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Task: {'id': task-1205724, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.30326} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.116129] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 793.116348] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 793.116533] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 793.127016] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Creating linked-clone VM from snapshot {{(pid=62385) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 793.127016] env[62385]: DEBUG oslo_vmware.api [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205728, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.128933] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-890dbb46-d243-407f-b8fa-def97e205e2d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.132827] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Created folder: Project (6e99f1e4cfd84bf083c8261b2ffcfa2f) in parent group-v261107. [ 793.132827] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Creating folder: Instances. Parent ref: group-v261184. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 793.132979] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cce72eec-5c66-4ac1-b9af-1751d5343f37 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.145024] env[62385]: DEBUG oslo_vmware.api [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Waiting for the task: (returnval){ [ 793.145024] env[62385]: value = "task-1205730" [ 793.145024] env[62385]: _type = "Task" [ 793.145024] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.153185] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Created folder: Instances in parent group-v261184. [ 793.153477] env[62385]: DEBUG oslo.service.loopingcall [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 793.158979] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 793.158979] env[62385]: DEBUG oslo_vmware.api [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205730, 'name': CloneVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.158979] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c2317e4e-a6b2-40d6-a9d3-4ffd336a8a61 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.184499] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 793.184499] env[62385]: value = "task-1205732" [ 793.184499] env[62385]: _type = "Task" [ 793.184499] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.193971] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205732, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.302929] env[62385]: DEBUG nova.network.neutron [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Updating instance_info_cache with network_info: [{"id": "854fc77a-662c-4fbe-a30c-5726d1ed6088", "address": "fa:16:3e:a6:e0:c0", "network": {"id": "b563b70c-d3f1-4f82-97e9-defaf3f9d14b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-55206131-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "8eaa81326f084441a87ac05796958abb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa", "external-id": "nsx-vlan-transportzone-132", "segmentation_id": 132, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap854fc77a-66", "ovs_interfaceid": "854fc77a-662c-4fbe-a30c-5726d1ed6088", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.366574] env[62385]: DEBUG oslo_vmware.api [None req-650314b1-bfcb-451d-9c85-40f84a6535ea tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Task: {'id': task-1205726, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.300617} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.366965] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-650314b1-bfcb-451d-9c85-40f84a6535ea tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 793.367192] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-650314b1-bfcb-451d-9c85-40f84a6535ea tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Deleted contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 793.367421] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-650314b1-bfcb-451d-9c85-40f84a6535ea tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 793.367616] env[62385]: INFO nova.compute.manager [None req-650314b1-bfcb-451d-9c85-40f84a6535ea tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Took 1.13 seconds to destroy the instance on the hypervisor. [ 793.367914] env[62385]: DEBUG oslo.service.loopingcall [None req-650314b1-bfcb-451d-9c85-40f84a6535ea tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 793.368162] env[62385]: DEBUG nova.compute.manager [-] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 793.368381] env[62385]: DEBUG nova.network.neutron [-] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 793.401867] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c3bb0442-402a-44c3-a5ec-63b1cb851d77 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "42ecf594-dff7-4af3-ac56-5bbc45b7a192" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.893s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.421675] env[62385]: DEBUG nova.compute.manager [req-52565f0c-ddba-4234-8f77-6cc795358b48 req-858d6dcd-8fc0-48c8-8200-a266f9327cce service nova] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Received event network-changed-2c31d161-814f-4c23-9882-0aa436bb9e03 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 793.422357] env[62385]: DEBUG nova.compute.manager [req-52565f0c-ddba-4234-8f77-6cc795358b48 req-858d6dcd-8fc0-48c8-8200-a266f9327cce service nova] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Refreshing instance network info cache due to event network-changed-2c31d161-814f-4c23-9882-0aa436bb9e03. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 793.422357] env[62385]: DEBUG oslo_concurrency.lockutils [req-52565f0c-ddba-4234-8f77-6cc795358b48 req-858d6dcd-8fc0-48c8-8200-a266f9327cce service nova] Acquiring lock "refresh_cache-d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.422357] env[62385]: DEBUG oslo_concurrency.lockutils [req-52565f0c-ddba-4234-8f77-6cc795358b48 req-858d6dcd-8fc0-48c8-8200-a266f9327cce service nova] Acquired lock "refresh_cache-d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.422772] env[62385]: DEBUG nova.network.neutron [req-52565f0c-ddba-4234-8f77-6cc795358b48 req-858d6dcd-8fc0-48c8-8200-a266f9327cce service nova] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Refreshing network info cache for port 2c31d161-814f-4c23-9882-0aa436bb9e03 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 793.525191] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.537719] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c91e46d6-47a6-401c-be1d-a84db176ef5a tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Lock "fd7a77e2-6d76-4d91-bdef-e30333247aa9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 96.975s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.613834] env[62385]: DEBUG oslo_vmware.api [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205728, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.188083} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.614113] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 793.614271] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Deleted contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 793.614483] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 793.633731] env[62385]: INFO nova.scheduler.client.report [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Deleted allocations for instance f8a98710-9d02-4aa3-b6a5-34b7060ad62d [ 793.656454] env[62385]: DEBUG oslo_vmware.api [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205730, 'name': CloneVM_Task} progress is 94%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.699149] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205732, 'name': CreateVM_Task, 'duration_secs': 0.398406} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.699335] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 793.700421] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.700658] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.700987] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 793.701285] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-40605709-9622-42d7-80b1-505938dbe908 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.709200] env[62385]: DEBUG oslo_vmware.api [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for the task: (returnval){ [ 793.709200] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]529fecb0-41ba-beef-8f18-325deeaac505" [ 793.709200] env[62385]: _type = "Task" [ 793.709200] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.718956] env[62385]: DEBUG oslo_vmware.api [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]529fecb0-41ba-beef-8f18-325deeaac505, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.760901] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8760ccbb-b3a3-4d00-86f3-b32f90b9d13b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.770019] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fc94595-f344-437d-8d6a-ac342a3e6b91 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.803213] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83beb6a6-c4ba-498c-bb02-44d5fcd23ccd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.806405] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Releasing lock "refresh_cache-b4970e1a-2b29-44e9-b79e-cda3c32dddcd" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.814346] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a70f057-2eae-4739-9010-2bdb328ba162 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.830097] env[62385]: DEBUG nova.compute.provider_tree [None req-0e43bfd4-0da1-49c6-a0cb-aa2b4014286e tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 794.042014] env[62385]: DEBUG nova.compute.manager [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 794.139960] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.162572] env[62385]: DEBUG oslo_vmware.api [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205730, 'name': CloneVM_Task} progress is 94%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.167895] env[62385]: DEBUG nova.virt.hardware [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 794.168272] env[62385]: DEBUG nova.virt.hardware [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 794.168765] env[62385]: DEBUG nova.virt.hardware [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 794.169067] env[62385]: DEBUG nova.virt.hardware [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 794.169341] env[62385]: DEBUG nova.virt.hardware [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 794.169566] env[62385]: DEBUG nova.virt.hardware [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 794.169816] env[62385]: DEBUG nova.virt.hardware [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 794.170018] env[62385]: DEBUG nova.virt.hardware [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 794.170237] env[62385]: DEBUG nova.virt.hardware [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 794.170430] env[62385]: DEBUG nova.virt.hardware [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 794.172594] env[62385]: DEBUG nova.virt.hardware [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 794.172594] env[62385]: DEBUG nova.network.neutron [-] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.173818] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-170ca15b-0e25-46e6-be4d-ae24431f9ec3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.185423] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da6fc81c-2256-4665-b07c-7e64d0c48c94 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.205097] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Instance VIF info [] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 794.214189] env[62385]: DEBUG oslo.service.loopingcall [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 794.214189] env[62385]: DEBUG nova.network.neutron [req-52565f0c-ddba-4234-8f77-6cc795358b48 req-858d6dcd-8fc0-48c8-8200-a266f9327cce service nova] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Updated VIF entry in instance network info cache for port 2c31d161-814f-4c23-9882-0aa436bb9e03. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 794.214524] env[62385]: DEBUG nova.network.neutron [req-52565f0c-ddba-4234-8f77-6cc795358b48 req-858d6dcd-8fc0-48c8-8200-a266f9327cce service nova] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Updating instance_info_cache with network_info: [{"id": "2c31d161-814f-4c23-9882-0aa436bb9e03", "address": "fa:16:3e:d7:43:42", "network": {"id": "aca904e1-fbd6-4def-bbec-bd70c0fc8af9", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-249813742-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e99f1e4cfd84bf083c8261b2ffcfa2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f7abe5d-b8fe-4983-bd50-e7469f1fe7f3", "external-id": "nsx-vlan-transportzone-263", "segmentation_id": 263, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c31d161-81", "ovs_interfaceid": "2c31d161-814f-4c23-9882-0aa436bb9e03", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.216163] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 794.219788] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bf7fc988-414c-465c-adb8-bcfb488febcf {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.237329] env[62385]: DEBUG nova.network.neutron [req-e8d29cdd-6b86-4836-b2b7-409cc1572187 req-872a7895-049b-4fef-a21d-5149d0f7ee8a service nova] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Updated VIF entry in instance network info cache for port bfb1f5cb-27f6-4422-8a40-f8f2d8a6520b. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 794.238852] env[62385]: DEBUG nova.network.neutron [req-e8d29cdd-6b86-4836-b2b7-409cc1572187 req-872a7895-049b-4fef-a21d-5149d0f7ee8a service nova] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Updating instance_info_cache with network_info: [{"id": "bfb1f5cb-27f6-4422-8a40-f8f2d8a6520b", "address": "fa:16:3e:0c:f4:9c", "network": {"id": "ff21dd88-fa02-4bd6-8009-980ec6e72758", "bridge": null, "label": "tempest-DeleteServersTestJSON-1065789169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aea48968d4464ac0ab01c998c059107d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapbfb1f5cb-27", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.243742] env[62385]: DEBUG oslo_vmware.api [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]529fecb0-41ba-beef-8f18-325deeaac505, 'name': SearchDatastore_Task, 'duration_secs': 0.01102} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.245459] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.245459] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 794.245669] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.245845] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.246188] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 794.246328] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 794.246328] env[62385]: value = "task-1205733" [ 794.246328] env[62385]: _type = "Task" [ 794.246328] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.246601] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-00f46ff6-e275-45ae-ba3b-6766027e22e7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.267758] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205733, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.277930] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 794.277930] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 794.279249] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f1b8effc-c790-4684-8b0a-55fcd49c9ff7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.286149] env[62385]: DEBUG oslo_vmware.api [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for the task: (returnval){ [ 794.286149] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52f09667-4790-e59c-ea13-03059a6c7004" [ 794.286149] env[62385]: _type = "Task" [ 794.286149] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.297397] env[62385]: DEBUG oslo_vmware.api [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52f09667-4790-e59c-ea13-03059a6c7004, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.333370] env[62385]: DEBUG nova.scheduler.client.report [None req-0e43bfd4-0da1-49c6-a0cb-aa2b4014286e tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 794.340123] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 794.340301] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bf24260f-b778-433d-b0aa-4ed87a751f20 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.348549] env[62385]: DEBUG oslo_vmware.api [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for the task: (returnval){ [ 794.348549] env[62385]: value = "task-1205734" [ 794.348549] env[62385]: _type = "Task" [ 794.348549] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.360897] env[62385]: DEBUG oslo_vmware.api [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205734, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.568815] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.660362] env[62385]: DEBUG oslo_vmware.api [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205730, 'name': CloneVM_Task} progress is 95%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.680053] env[62385]: INFO nova.compute.manager [-] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Took 1.31 seconds to deallocate network for instance. [ 794.718100] env[62385]: DEBUG oslo_concurrency.lockutils [req-52565f0c-ddba-4234-8f77-6cc795358b48 req-858d6dcd-8fc0-48c8-8200-a266f9327cce service nova] Releasing lock "refresh_cache-d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.744733] env[62385]: DEBUG oslo_concurrency.lockutils [req-e8d29cdd-6b86-4836-b2b7-409cc1572187 req-872a7895-049b-4fef-a21d-5149d0f7ee8a service nova] Releasing lock "refresh_cache-f8a98710-9d02-4aa3-b6a5-34b7060ad62d" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.759896] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205733, 'name': CreateVM_Task, 'duration_secs': 0.355053} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.760147] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 794.760664] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.760877] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.761387] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 794.761682] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b298825-36d6-45b9-ba61-c1a9f254bc9f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.767853] env[62385]: DEBUG oslo_vmware.api [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Waiting for the task: (returnval){ [ 794.767853] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e46ecc-5b1c-81c0-ea05-847f37867d9a" [ 794.767853] env[62385]: _type = "Task" [ 794.767853] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.775717] env[62385]: DEBUG oslo_vmware.api [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e46ecc-5b1c-81c0-ea05-847f37867d9a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.797425] env[62385]: DEBUG oslo_vmware.api [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52f09667-4790-e59c-ea13-03059a6c7004, 'name': SearchDatastore_Task, 'duration_secs': 0.013419} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.798338] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8de526e6-0434-4f0c-b5a2-d488c9c992a6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.805081] env[62385]: DEBUG oslo_vmware.api [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for the task: (returnval){ [ 794.805081] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c123bb-a53e-ecca-13ff-77e8d090f130" [ 794.805081] env[62385]: _type = "Task" [ 794.805081] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.815038] env[62385]: DEBUG oslo_vmware.api [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c123bb-a53e-ecca-13ff-77e8d090f130, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.838421] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0e43bfd4-0da1-49c6-a0cb-aa2b4014286e tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.972s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.843307] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 29.374s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.843510] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.843676] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62385) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 794.843983] env[62385]: DEBUG oslo_concurrency.lockutils [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.523s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.845612] env[62385]: INFO nova.compute.claims [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 794.849969] env[62385]: DEBUG oslo_concurrency.lockutils [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "b73a31df-53c8-4550-bf75-5cf3b5aff86c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.852455] env[62385]: DEBUG oslo_concurrency.lockutils [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "b73a31df-53c8-4550-bf75-5cf3b5aff86c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.852455] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd34c85d-1c86-4f02-85a5-6ab0b48c0cb5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.862084] env[62385]: INFO nova.scheduler.client.report [None req-0e43bfd4-0da1-49c6-a0cb-aa2b4014286e tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Deleted allocations for instance 92057af7-28a3-4643-9cda-d3d868d0cfc9 [ 794.876838] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99129902-397d-4ae9-8f68-f6bd8d2a3945 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.885518] env[62385]: DEBUG oslo_vmware.api [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205734, 'name': PowerOffVM_Task, 'duration_secs': 0.262496} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.886700] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 794.888661] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64e9b937-e621-48a8-a320-b3e59d000521 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.905782] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a00092d4-52af-428f-b8df-2fb2c0f98e67 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.923786] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0015eb7-d683-433a-825e-eed9290336e0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.930060] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9190ec18-db23-4a91-8fe7-7b627f61dc55 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.967115] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181003MB free_disk=155GB free_vcpus=48 pci_devices=None {{(pid=62385) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 794.967346] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.976010] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 794.976338] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5fb5a5e3-0972-4c07-a5bb-67126560b80f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.986172] env[62385]: DEBUG oslo_vmware.api [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for the task: (returnval){ [ 794.986172] env[62385]: value = "task-1205735" [ 794.986172] env[62385]: _type = "Task" [ 794.986172] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.995868] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] VM already powered off {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 794.996098] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 794.996354] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.158032] env[62385]: DEBUG oslo_vmware.api [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205730, 'name': CloneVM_Task, 'duration_secs': 1.786109} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.158395] env[62385]: INFO nova.virt.vmwareapi.vmops [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Created linked-clone VM from snapshot [ 795.159166] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ebd11de-c66d-40c6-b743-375e22213123 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.167446] env[62385]: DEBUG nova.virt.vmwareapi.images [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Uploading image 28813632-55e0-4297-8acd-1f1992ee8fe5 {{(pid=62385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 795.186928] env[62385]: DEBUG oslo_vmware.rw_handles [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 795.186928] env[62385]: value = "vm-261186" [ 795.186928] env[62385]: _type = "VirtualMachine" [ 795.186928] env[62385]: }. {{(pid=62385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 795.187793] env[62385]: DEBUG oslo_concurrency.lockutils [None req-650314b1-bfcb-451d-9c85-40f84a6535ea tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.188449] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-7e822bcb-80b4-451b-82a9-314f97659b8a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.196531] env[62385]: DEBUG oslo_vmware.rw_handles [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Lease: (returnval){ [ 795.196531] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521fbd16-8f78-f996-870c-836f91e0cc2d" [ 795.196531] env[62385]: _type = "HttpNfcLease" [ 795.196531] env[62385]: } obtained for exporting VM: (result){ [ 795.196531] env[62385]: value = "vm-261186" [ 795.196531] env[62385]: _type = "VirtualMachine" [ 795.196531] env[62385]: }. {{(pid=62385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 795.196790] env[62385]: DEBUG oslo_vmware.api [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Waiting for the lease: (returnval){ [ 795.196790] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521fbd16-8f78-f996-870c-836f91e0cc2d" [ 795.196790] env[62385]: _type = "HttpNfcLease" [ 795.196790] env[62385]: } to be ready. {{(pid=62385) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 795.204380] env[62385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 795.204380] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521fbd16-8f78-f996-870c-836f91e0cc2d" [ 795.204380] env[62385]: _type = "HttpNfcLease" [ 795.204380] env[62385]: } is initializing. {{(pid=62385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 795.285261] env[62385]: DEBUG oslo_vmware.api [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e46ecc-5b1c-81c0-ea05-847f37867d9a, 'name': SearchDatastore_Task, 'duration_secs': 0.012537} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.285492] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.285741] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 795.285977] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.286171] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.286381] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 795.286659] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1027c42a-16f3-4d39-b909-ec5bae1195fb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.297063] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 795.297291] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Folder [datastore2] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 795.298036] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c88c1619-5c41-4d36-b88e-1cf18a1ade53 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.304901] env[62385]: DEBUG oslo_vmware.api [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Waiting for the task: (returnval){ [ 795.304901] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5207fea4-d107-4492-10d7-fedba5dd3458" [ 795.304901] env[62385]: _type = "Task" [ 795.304901] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.317345] env[62385]: DEBUG oslo_vmware.api [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5207fea4-d107-4492-10d7-fedba5dd3458, 'name': SearchDatastore_Task, 'duration_secs': 0.009476} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.321365] env[62385]: DEBUG oslo_vmware.api [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c123bb-a53e-ecca-13ff-77e8d090f130, 'name': SearchDatastore_Task, 'duration_secs': 0.013218} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.321450] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f69e7332-db81-4a62-a7d1-46c361bdc448 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.323537] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.323801] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b/d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 795.324088] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.324282] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 795.324484] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8ded967e-d1d4-4f9b-83d0-3ec991bac70d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.326766] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fdf69ecf-c4dc-4761-b76a-f49a71c36e79 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.330241] env[62385]: DEBUG oslo_vmware.api [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Waiting for the task: (returnval){ [ 795.330241] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521959ba-d15d-81e6-c38b-1ab587e47588" [ 795.330241] env[62385]: _type = "Task" [ 795.330241] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.335048] env[62385]: DEBUG oslo_vmware.api [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for the task: (returnval){ [ 795.335048] env[62385]: value = "task-1205737" [ 795.335048] env[62385]: _type = "Task" [ 795.335048] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.341666] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 795.341849] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 795.342579] env[62385]: DEBUG oslo_vmware.api [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521959ba-d15d-81e6-c38b-1ab587e47588, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.343149] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5eed5a41-367f-4f59-b1ee-2d28669b736f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.348129] env[62385]: DEBUG oslo_vmware.api [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1205737, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.353374] env[62385]: DEBUG oslo_vmware.api [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for the task: (returnval){ [ 795.353374] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]526b6a56-bccf-c371-b65b-ec33fa62807e" [ 795.353374] env[62385]: _type = "Task" [ 795.353374] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.362699] env[62385]: DEBUG oslo_vmware.api [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]526b6a56-bccf-c371-b65b-ec33fa62807e, 'name': SearchDatastore_Task} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.363657] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96d51a99-268b-4161-86bb-4674c4169bd4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.372313] env[62385]: DEBUG oslo_vmware.api [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for the task: (returnval){ [ 795.372313] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ad409b-57fc-7933-8dde-0f4722b6a66a" [ 795.372313] env[62385]: _type = "Task" [ 795.372313] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.372751] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0e43bfd4-0da1-49c6-a0cb-aa2b4014286e tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Lock "92057af7-28a3-4643-9cda-d3d868d0cfc9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.717s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.383365] env[62385]: DEBUG oslo_vmware.api [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ad409b-57fc-7933-8dde-0f4722b6a66a, 'name': SearchDatastore_Task, 'duration_secs': 0.010892} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.383695] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.383926] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] b4970e1a-2b29-44e9-b79e-cda3c32dddcd/c0abbb8d-77e7-4f22-a256-0faf0a781109-rescue.vmdk. {{(pid=62385) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 795.384223] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e2026ec9-6bd9-45e6-b79e-353a27dea4a5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.393203] env[62385]: DEBUG oslo_vmware.api [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for the task: (returnval){ [ 795.393203] env[62385]: value = "task-1205738" [ 795.393203] env[62385]: _type = "Task" [ 795.393203] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.403387] env[62385]: DEBUG oslo_vmware.api [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205738, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.547338] env[62385]: DEBUG nova.compute.manager [req-2429b4f2-616a-45bd-98e9-1441b34c2189 req-492e4ce7-3e3a-460c-905a-bd80ef26246d service nova] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Received event network-vif-deleted-b838271e-07e8-4880-980a-703c1bbd4da5 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 795.558762] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3e237177-68bb-4b96-9885-fed59468a6c1 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "f8a98710-9d02-4aa3-b6a5-34b7060ad62d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.707423] env[62385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 795.707423] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521fbd16-8f78-f996-870c-836f91e0cc2d" [ 795.707423] env[62385]: _type = "HttpNfcLease" [ 795.707423] env[62385]: } is ready. {{(pid=62385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 795.707824] env[62385]: DEBUG oslo_vmware.rw_handles [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 795.707824] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521fbd16-8f78-f996-870c-836f91e0cc2d" [ 795.707824] env[62385]: _type = "HttpNfcLease" [ 795.707824] env[62385]: }. {{(pid=62385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 795.708840] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffdcc403-23f3-41df-987b-5896cf266d01 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.718436] env[62385]: DEBUG oslo_vmware.rw_handles [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5255f5df-ab9a-92a7-36ec-af386ab79ff8/disk-0.vmdk from lease info. {{(pid=62385) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 795.719845] env[62385]: DEBUG oslo_vmware.rw_handles [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5255f5df-ab9a-92a7-36ec-af386ab79ff8/disk-0.vmdk for reading. {{(pid=62385) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 795.844822] env[62385]: DEBUG oslo_vmware.api [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521959ba-d15d-81e6-c38b-1ab587e47588, 'name': SearchDatastore_Task, 'duration_secs': 0.0097} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.844822] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.845064] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7/8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 795.845350] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ee5c2e37-b83f-48ad-9b23-59cc45ffa128 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.850849] env[62385]: DEBUG oslo_vmware.api [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1205737, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.861305] env[62385]: DEBUG oslo_vmware.api [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Waiting for the task: (returnval){ [ 795.861305] env[62385]: value = "task-1205739" [ 795.861305] env[62385]: _type = "Task" [ 795.861305] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.871442] env[62385]: DEBUG oslo_vmware.api [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Task: {'id': task-1205739, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.895633] env[62385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-d8c1ee83-67d4-48a9-b6e2-4ee0ef42ed91 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.912888] env[62385]: DEBUG oslo_vmware.api [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205738, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.291719] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10e66147-44fa-4d5d-a1a1-6ae663c2a1e5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.300146] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ee999fe-1f7b-4c56-b15f-34e42f0113e4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.332747] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97170634-43a8-44ba-b68f-a2401a07c9ab {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.345414] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-008a99ae-f268-47c4-ad7a-b262f5891f9e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.353201] env[62385]: DEBUG oslo_vmware.api [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1205737, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.547167} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.354200] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b/d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 796.355132] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 796.355132] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-69eb0902-1426-4d38-9aaa-c01f544181c4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.365369] env[62385]: DEBUG nova.compute.provider_tree [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 796.374522] env[62385]: DEBUG oslo_vmware.api [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for the task: (returnval){ [ 796.374522] env[62385]: value = "task-1205740" [ 796.374522] env[62385]: _type = "Task" [ 796.374522] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.378027] env[62385]: DEBUG oslo_vmware.api [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Task: {'id': task-1205739, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.389294] env[62385]: DEBUG oslo_vmware.api [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1205740, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.406501] env[62385]: DEBUG oslo_vmware.api [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205738, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.847594} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.407076] env[62385]: INFO nova.virt.vmwareapi.ds_util [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] b4970e1a-2b29-44e9-b79e-cda3c32dddcd/c0abbb8d-77e7-4f22-a256-0faf0a781109-rescue.vmdk. [ 796.407974] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-897960c0-cc45-4bc8-905a-01a74c7c80d8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.437628] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Reconfiguring VM instance instance-0000003c to attach disk [datastore1] b4970e1a-2b29-44e9-b79e-cda3c32dddcd/c0abbb8d-77e7-4f22-a256-0faf0a781109-rescue.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 796.438617] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-92bac55b-6ade-4e62-a1e9-4f357df0e5d0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.462451] env[62385]: DEBUG oslo_vmware.api [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for the task: (returnval){ [ 796.462451] env[62385]: value = "task-1205741" [ 796.462451] env[62385]: _type = "Task" [ 796.462451] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.472233] env[62385]: DEBUG oslo_vmware.api [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205741, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.875149] env[62385]: DEBUG nova.scheduler.client.report [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 796.879994] env[62385]: DEBUG oslo_vmware.api [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Task: {'id': task-1205739, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.823845} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.879994] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7/8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 796.879994] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 796.882725] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-560920e9-7fc5-4cf8-bc99-048e66057a46 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.891519] env[62385]: DEBUG oslo_vmware.api [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1205740, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.090671} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.892920] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 796.893361] env[62385]: DEBUG oslo_vmware.api [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Waiting for the task: (returnval){ [ 796.893361] env[62385]: value = "task-1205742" [ 796.893361] env[62385]: _type = "Task" [ 796.893361] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.894931] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1cc5f28-f3e0-4328-9576-fdbd6c914d12 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.909823] env[62385]: DEBUG oslo_vmware.api [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Task: {'id': task-1205742, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.927925] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Reconfiguring VM instance instance-0000003e to attach disk [datastore1] d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b/d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 796.928402] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b873c53e-b7ae-4fa5-bd8a-8f8f7affee8d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.951802] env[62385]: DEBUG oslo_vmware.api [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for the task: (returnval){ [ 796.951802] env[62385]: value = "task-1205743" [ 796.951802] env[62385]: _type = "Task" [ 796.951802] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.961764] env[62385]: DEBUG oslo_vmware.api [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1205743, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.972554] env[62385]: DEBUG oslo_vmware.api [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205741, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.209250] env[62385]: DEBUG oslo_concurrency.lockutils [None req-aedae9b2-11c8-443b-9154-3314c6ae37ea tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Acquiring lock "fd7a77e2-6d76-4d91-bdef-e30333247aa9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.209644] env[62385]: DEBUG oslo_concurrency.lockutils [None req-aedae9b2-11c8-443b-9154-3314c6ae37ea tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Lock "fd7a77e2-6d76-4d91-bdef-e30333247aa9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.209939] env[62385]: DEBUG oslo_concurrency.lockutils [None req-aedae9b2-11c8-443b-9154-3314c6ae37ea tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Acquiring lock "fd7a77e2-6d76-4d91-bdef-e30333247aa9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.210269] env[62385]: DEBUG oslo_concurrency.lockutils [None req-aedae9b2-11c8-443b-9154-3314c6ae37ea tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Lock "fd7a77e2-6d76-4d91-bdef-e30333247aa9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.210497] env[62385]: DEBUG oslo_concurrency.lockutils [None req-aedae9b2-11c8-443b-9154-3314c6ae37ea tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Lock "fd7a77e2-6d76-4d91-bdef-e30333247aa9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.213157] env[62385]: INFO nova.compute.manager [None req-aedae9b2-11c8-443b-9154-3314c6ae37ea tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Terminating instance [ 797.216995] env[62385]: DEBUG nova.compute.manager [None req-aedae9b2-11c8-443b-9154-3314c6ae37ea tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 797.217223] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-aedae9b2-11c8-443b-9154-3314c6ae37ea tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 797.218240] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc81990e-b515-4967-b0a5-78ae6267cf3d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.229083] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-aedae9b2-11c8-443b-9154-3314c6ae37ea tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 797.229083] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b5494945-f8ad-4a3e-89d2-98e30b326556 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.237521] env[62385]: DEBUG oslo_vmware.api [None req-aedae9b2-11c8-443b-9154-3314c6ae37ea tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Waiting for the task: (returnval){ [ 797.237521] env[62385]: value = "task-1205744" [ 797.237521] env[62385]: _type = "Task" [ 797.237521] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.247507] env[62385]: DEBUG oslo_vmware.api [None req-aedae9b2-11c8-443b-9154-3314c6ae37ea tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Task: {'id': task-1205744, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.260268] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Acquiring lock "2a579d13-5372-4340-b7b3-cc02c1912624" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.260623] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Lock "2a579d13-5372-4340-b7b3-cc02c1912624" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.380806] env[62385]: DEBUG oslo_concurrency.lockutils [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.536s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.381776] env[62385]: DEBUG nova.compute.manager [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 797.386744] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.002s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.388474] env[62385]: INFO nova.compute.claims [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] [instance: 2b744879-01d2-440e-8adf-58438b922888] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 797.407720] env[62385]: DEBUG oslo_vmware.api [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Task: {'id': task-1205742, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068207} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.408139] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 797.409153] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47bbf0d0-1018-4e74-b2c1-2835c8005a1a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.431391] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Reconfiguring VM instance instance-0000003a to attach disk [datastore2] 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7/8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 797.433559] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f28aa721-6bc7-4fae-afe1-580c57684b81 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.458704] env[62385]: DEBUG oslo_vmware.api [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Waiting for the task: (returnval){ [ 797.458704] env[62385]: value = "task-1205745" [ 797.458704] env[62385]: _type = "Task" [ 797.458704] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.466587] env[62385]: DEBUG oslo_vmware.api [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1205743, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.474627] env[62385]: DEBUG oslo_vmware.api [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Task: {'id': task-1205745, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.480215] env[62385]: DEBUG oslo_vmware.api [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205741, 'name': ReconfigVM_Task, 'duration_secs': 0.725568} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.480803] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Reconfigured VM instance instance-0000003c to attach disk [datastore1] b4970e1a-2b29-44e9-b79e-cda3c32dddcd/c0abbb8d-77e7-4f22-a256-0faf0a781109-rescue.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 797.481806] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be1d7404-0e1f-4298-83f6-bbe8804de7b8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.511955] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e7569c1e-314a-4907-aedd-0b1fdefb8c7a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.530830] env[62385]: DEBUG oslo_vmware.api [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for the task: (returnval){ [ 797.530830] env[62385]: value = "task-1205746" [ 797.530830] env[62385]: _type = "Task" [ 797.530830] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.541075] env[62385]: DEBUG oslo_vmware.api [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205746, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.747530] env[62385]: DEBUG oslo_vmware.api [None req-aedae9b2-11c8-443b-9154-3314c6ae37ea tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Task: {'id': task-1205744, 'name': PowerOffVM_Task, 'duration_secs': 0.205816} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.747873] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-aedae9b2-11c8-443b-9154-3314c6ae37ea tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 797.748142] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-aedae9b2-11c8-443b-9154-3314c6ae37ea tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 797.748493] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d50a9baf-016d-4a2e-9dd8-204eed9e060c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.820175] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-aedae9b2-11c8-443b-9154-3314c6ae37ea tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 797.820441] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-aedae9b2-11c8-443b-9154-3314c6ae37ea tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 797.820734] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-aedae9b2-11c8-443b-9154-3314c6ae37ea tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Deleting the datastore file [datastore1] fd7a77e2-6d76-4d91-bdef-e30333247aa9 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 797.821117] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3ebd4bb2-3e27-4c20-8870-b4ac78878abb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.829494] env[62385]: DEBUG oslo_vmware.api [None req-aedae9b2-11c8-443b-9154-3314c6ae37ea tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Waiting for the task: (returnval){ [ 797.829494] env[62385]: value = "task-1205748" [ 797.829494] env[62385]: _type = "Task" [ 797.829494] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.841243] env[62385]: DEBUG oslo_vmware.api [None req-aedae9b2-11c8-443b-9154-3314c6ae37ea tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Task: {'id': task-1205748, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.894413] env[62385]: DEBUG nova.compute.utils [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 797.900013] env[62385]: DEBUG nova.compute.manager [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 797.900247] env[62385]: DEBUG nova.network.neutron [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 797.961540] env[62385]: DEBUG nova.policy [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b314c392b9ca405480c27c25364fda81', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '93111f5ba17c47de9691f90b99e03aa8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 797.970065] env[62385]: DEBUG oslo_vmware.api [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1205743, 'name': ReconfigVM_Task, 'duration_secs': 0.563092} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.970497] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Reconfigured VM instance instance-0000003e to attach disk [datastore1] d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b/d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 797.971759] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e1bbdcf2-6d2b-43ea-8ba6-2238fed6a96f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.977238] env[62385]: DEBUG oslo_vmware.api [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Task: {'id': task-1205745, 'name': ReconfigVM_Task, 'duration_secs': 0.285282} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.977567] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Reconfigured VM instance instance-0000003a to attach disk [datastore2] 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7/8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 797.978158] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c7cb5c64-e5b0-4745-a5d2-f69bdc09407e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.982441] env[62385]: DEBUG oslo_vmware.api [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for the task: (returnval){ [ 797.982441] env[62385]: value = "task-1205749" [ 797.982441] env[62385]: _type = "Task" [ 797.982441] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.987084] env[62385]: DEBUG oslo_vmware.api [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Waiting for the task: (returnval){ [ 797.987084] env[62385]: value = "task-1205750" [ 797.987084] env[62385]: _type = "Task" [ 797.987084] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.994151] env[62385]: DEBUG oslo_vmware.api [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1205749, 'name': Rename_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.999795] env[62385]: DEBUG oslo_vmware.api [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Task: {'id': task-1205750, 'name': Rename_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.043554] env[62385]: DEBUG oslo_vmware.api [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205746, 'name': ReconfigVM_Task, 'duration_secs': 0.160732} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.044065] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 798.044392] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ac2ac9e1-9dfa-48c9-86e7-51cf32685ea4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.055051] env[62385]: DEBUG oslo_vmware.api [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for the task: (returnval){ [ 798.055051] env[62385]: value = "task-1205751" [ 798.055051] env[62385]: _type = "Task" [ 798.055051] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.340804] env[62385]: DEBUG oslo_vmware.api [None req-aedae9b2-11c8-443b-9154-3314c6ae37ea tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Task: {'id': task-1205748, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.247156} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.341544] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-aedae9b2-11c8-443b-9154-3314c6ae37ea tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 798.341544] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-aedae9b2-11c8-443b-9154-3314c6ae37ea tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 798.341544] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-aedae9b2-11c8-443b-9154-3314c6ae37ea tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 798.341811] env[62385]: INFO nova.compute.manager [None req-aedae9b2-11c8-443b-9154-3314c6ae37ea tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Took 1.12 seconds to destroy the instance on the hypervisor. [ 798.341884] env[62385]: DEBUG oslo.service.loopingcall [None req-aedae9b2-11c8-443b-9154-3314c6ae37ea tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 798.342103] env[62385]: DEBUG nova.compute.manager [-] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 798.342204] env[62385]: DEBUG nova.network.neutron [-] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 798.402851] env[62385]: DEBUG nova.compute.manager [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 798.497539] env[62385]: DEBUG oslo_vmware.api [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1205749, 'name': Rename_Task, 'duration_secs': 0.157711} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.497836] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 798.498528] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5c1f021a-ba93-453c-a34a-1139b97a56e8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.507560] env[62385]: DEBUG oslo_vmware.api [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Task: {'id': task-1205750, 'name': Rename_Task, 'duration_secs': 0.164185} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.507875] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 798.508213] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-28ecb067-c2f7-46f2-a1c8-e6a8c7cf1a0a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.513341] env[62385]: DEBUG oslo_vmware.api [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for the task: (returnval){ [ 798.513341] env[62385]: value = "task-1205752" [ 798.513341] env[62385]: _type = "Task" [ 798.513341] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.520491] env[62385]: DEBUG oslo_vmware.api [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Waiting for the task: (returnval){ [ 798.520491] env[62385]: value = "task-1205753" [ 798.520491] env[62385]: _type = "Task" [ 798.520491] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.530436] env[62385]: DEBUG oslo_vmware.api [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1205752, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.536959] env[62385]: DEBUG oslo_vmware.api [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Task: {'id': task-1205753, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.570159] env[62385]: DEBUG oslo_vmware.api [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205751, 'name': PowerOnVM_Task, 'duration_secs': 0.467649} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.570760] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 798.574560] env[62385]: DEBUG nova.compute.manager [None req-bfb8f513-8fe9-4e32-8cdd-c09f5e5566d4 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 798.576361] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5a3548d-8a37-4274-bc17-64651afae50c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.596026] env[62385]: DEBUG nova.network.neutron [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Successfully created port: ff03b686-6596-418e-b748-0a8a6799a2b5 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 798.673689] env[62385]: DEBUG nova.compute.manager [req-c3fab560-7409-489e-9643-c8cf457f25a8 req-34257454-387c-49b9-bacd-f470005ee48b service nova] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Received event network-vif-deleted-c14a2370-3801-4b30-ad4f-e86a6b5c133d {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 798.673916] env[62385]: INFO nova.compute.manager [req-c3fab560-7409-489e-9643-c8cf457f25a8 req-34257454-387c-49b9-bacd-f470005ee48b service nova] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Neutron deleted interface c14a2370-3801-4b30-ad4f-e86a6b5c133d; detaching it from the instance and deleting it from the info cache [ 798.674089] env[62385]: DEBUG nova.network.neutron [req-c3fab560-7409-489e-9643-c8cf457f25a8 req-34257454-387c-49b9-bacd-f470005ee48b service nova] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.821061] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb6fa618-87ce-4967-9c15-8b583edc138a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.830331] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4d1143c-5595-4083-9584-6700218b9f86 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.864508] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a13fd9f-c9f1-4182-a86a-70b4ce34b84c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.873715] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce5e6cf2-626b-4116-8cff-142dfa36e531 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.888515] env[62385]: DEBUG nova.compute.provider_tree [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 799.023802] env[62385]: DEBUG oslo_vmware.api [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1205752, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.034165] env[62385]: DEBUG oslo_vmware.api [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Task: {'id': task-1205753, 'name': PowerOnVM_Task, 'duration_secs': 0.450502} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.034165] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 799.034356] env[62385]: DEBUG nova.compute.manager [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 799.035292] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77f41fd8-0c55-4b43-af8c-150b753b201b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.146942] env[62385]: DEBUG nova.network.neutron [-] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.177207] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d1de2899-42c6-44d5-90dc-a61defa5b1dd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.188753] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-010249ff-d863-416a-8332-16d9decb4cd5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.220576] env[62385]: DEBUG nova.compute.manager [req-c3fab560-7409-489e-9643-c8cf457f25a8 req-34257454-387c-49b9-bacd-f470005ee48b service nova] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Detach interface failed, port_id=c14a2370-3801-4b30-ad4f-e86a6b5c133d, reason: Instance fd7a77e2-6d76-4d91-bdef-e30333247aa9 could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 799.395033] env[62385]: DEBUG nova.scheduler.client.report [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 799.416297] env[62385]: DEBUG nova.compute.manager [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 799.441749] env[62385]: DEBUG nova.virt.hardware [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 799.442015] env[62385]: DEBUG nova.virt.hardware [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 799.442900] env[62385]: DEBUG nova.virt.hardware [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 799.442900] env[62385]: DEBUG nova.virt.hardware [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 799.442900] env[62385]: DEBUG nova.virt.hardware [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 799.442900] env[62385]: DEBUG nova.virt.hardware [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 799.443092] env[62385]: DEBUG nova.virt.hardware [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 799.443190] env[62385]: DEBUG nova.virt.hardware [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 799.443366] env[62385]: DEBUG nova.virt.hardware [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 799.443538] env[62385]: DEBUG nova.virt.hardware [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 799.443717] env[62385]: DEBUG nova.virt.hardware [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 799.444993] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8fb8576-79f8-4ee1-80fc-3e134ffccfbb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.456465] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-904745aa-f65e-42f0-8a4f-8fe12292160a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.524049] env[62385]: DEBUG oslo_vmware.api [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1205752, 'name': PowerOnVM_Task, 'duration_secs': 0.658545} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.524341] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 799.524552] env[62385]: INFO nova.compute.manager [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Took 8.69 seconds to spawn the instance on the hypervisor. [ 799.524744] env[62385]: DEBUG nova.compute.manager [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 799.525547] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a222895-5563-42a7-a36f-4e064022e6f4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.558206] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.650848] env[62385]: INFO nova.compute.manager [-] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Took 1.31 seconds to deallocate network for instance. [ 799.900637] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.514s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.901212] env[62385]: DEBUG nova.compute.manager [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] [instance: 2b744879-01d2-440e-8adf-58438b922888] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 799.904730] env[62385]: DEBUG oslo_concurrency.lockutils [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.621s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.905843] env[62385]: INFO nova.compute.claims [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 799.975472] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2fbab2d2-44d4-4ed4-9df6-51fefeaa7131 tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Acquiring lock "8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.975831] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2fbab2d2-44d4-4ed4-9df6-51fefeaa7131 tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Lock "8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.976075] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2fbab2d2-44d4-4ed4-9df6-51fefeaa7131 tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Acquiring lock "8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.976312] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2fbab2d2-44d4-4ed4-9df6-51fefeaa7131 tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Lock "8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.976496] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2fbab2d2-44d4-4ed4-9df6-51fefeaa7131 tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Lock "8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.978837] env[62385]: INFO nova.compute.manager [None req-2fbab2d2-44d4-4ed4-9df6-51fefeaa7131 tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Terminating instance [ 799.980687] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2fbab2d2-44d4-4ed4-9df6-51fefeaa7131 tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Acquiring lock "refresh_cache-8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.980855] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2fbab2d2-44d4-4ed4-9df6-51fefeaa7131 tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Acquired lock "refresh_cache-8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.981057] env[62385]: DEBUG nova.network.neutron [None req-2fbab2d2-44d4-4ed4-9df6-51fefeaa7131 tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 800.050244] env[62385]: INFO nova.compute.manager [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Took 43.54 seconds to build instance. [ 800.158896] env[62385]: DEBUG oslo_concurrency.lockutils [None req-aedae9b2-11c8-443b-9154-3314c6ae37ea tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.322082] env[62385]: DEBUG nova.compute.manager [req-fb4dc771-2326-43ec-a43e-7b2a509dded5 req-1a2e530f-0129-41c4-9bea-4288c39e7cc7 service nova] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Received event network-vif-plugged-ff03b686-6596-418e-b748-0a8a6799a2b5 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 800.322337] env[62385]: DEBUG oslo_concurrency.lockutils [req-fb4dc771-2326-43ec-a43e-7b2a509dded5 req-1a2e530f-0129-41c4-9bea-4288c39e7cc7 service nova] Acquiring lock "9d6f098a-0b05-43ef-96b0-9eb99ad3538c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.322669] env[62385]: DEBUG oslo_concurrency.lockutils [req-fb4dc771-2326-43ec-a43e-7b2a509dded5 req-1a2e530f-0129-41c4-9bea-4288c39e7cc7 service nova] Lock "9d6f098a-0b05-43ef-96b0-9eb99ad3538c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.322934] env[62385]: DEBUG oslo_concurrency.lockutils [req-fb4dc771-2326-43ec-a43e-7b2a509dded5 req-1a2e530f-0129-41c4-9bea-4288c39e7cc7 service nova] Lock "9d6f098a-0b05-43ef-96b0-9eb99ad3538c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.323190] env[62385]: DEBUG nova.compute.manager [req-fb4dc771-2326-43ec-a43e-7b2a509dded5 req-1a2e530f-0129-41c4-9bea-4288c39e7cc7 service nova] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] No waiting events found dispatching network-vif-plugged-ff03b686-6596-418e-b748-0a8a6799a2b5 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 800.323484] env[62385]: WARNING nova.compute.manager [req-fb4dc771-2326-43ec-a43e-7b2a509dded5 req-1a2e530f-0129-41c4-9bea-4288c39e7cc7 service nova] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Received unexpected event network-vif-plugged-ff03b686-6596-418e-b748-0a8a6799a2b5 for instance with vm_state building and task_state spawning. [ 800.342992] env[62385]: DEBUG nova.network.neutron [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Successfully updated port: ff03b686-6596-418e-b748-0a8a6799a2b5 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 800.411525] env[62385]: DEBUG nova.compute.utils [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 800.414243] env[62385]: DEBUG nova.compute.manager [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] [instance: 2b744879-01d2-440e-8adf-58438b922888] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 800.414459] env[62385]: DEBUG nova.network.neutron [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] [instance: 2b744879-01d2-440e-8adf-58438b922888] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 800.498188] env[62385]: DEBUG nova.policy [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'eb0c585d7c254cdaab89c8fe0fb9657c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '073bfa66c3924f9db4f43ce2efff97b2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 800.502909] env[62385]: DEBUG nova.network.neutron [None req-2fbab2d2-44d4-4ed4-9df6-51fefeaa7131 tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 800.555022] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51ac681f-96d2-4cb0-a24e-d36aff81ed69 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 99.206s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.571749] env[62385]: DEBUG nova.network.neutron [None req-2fbab2d2-44d4-4ed4-9df6-51fefeaa7131 tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 800.625797] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Acquiring lock "5e5f43e8-ce45-45c7-a244-bfa724ea5e3c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.626160] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Lock "5e5f43e8-ce45-45c7-a244-bfa724ea5e3c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.851942] env[62385]: DEBUG oslo_concurrency.lockutils [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "refresh_cache-9d6f098a-0b05-43ef-96b0-9eb99ad3538c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.852218] env[62385]: DEBUG oslo_concurrency.lockutils [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquired lock "refresh_cache-9d6f098a-0b05-43ef-96b0-9eb99ad3538c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.852477] env[62385]: DEBUG nova.network.neutron [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 800.914992] env[62385]: DEBUG nova.compute.manager [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] [instance: 2b744879-01d2-440e-8adf-58438b922888] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 801.058056] env[62385]: DEBUG nova.compute.manager [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 801.073990] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2fbab2d2-44d4-4ed4-9df6-51fefeaa7131 tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Releasing lock "refresh_cache-8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.074466] env[62385]: DEBUG nova.compute.manager [None req-2fbab2d2-44d4-4ed4-9df6-51fefeaa7131 tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 801.074680] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-2fbab2d2-44d4-4ed4-9df6-51fefeaa7131 tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 801.075911] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29ba7f89-4135-4cee-8037-9ccc54fe6010 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.086255] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fbab2d2-44d4-4ed4-9df6-51fefeaa7131 tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 801.086377] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-971cc718-62f0-49ff-b48d-96b70deb1647 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.094780] env[62385]: DEBUG oslo_vmware.api [None req-2fbab2d2-44d4-4ed4-9df6-51fefeaa7131 tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Waiting for the task: (returnval){ [ 801.094780] env[62385]: value = "task-1205754" [ 801.094780] env[62385]: _type = "Task" [ 801.094780] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.108912] env[62385]: DEBUG oslo_vmware.api [None req-2fbab2d2-44d4-4ed4-9df6-51fefeaa7131 tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': task-1205754, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.223666] env[62385]: DEBUG nova.network.neutron [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] [instance: 2b744879-01d2-440e-8adf-58438b922888] Successfully created port: f42dfa8b-5ef4-40f7-9359-aa425ae2fbad {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 801.341601] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6bfacfd-22f0-4ed2-ad2f-53559c9734ff {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.350087] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e556f592-69ed-49a2-a04c-0cc2fad6594d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.382773] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a86d12b8-5ae5-4b1c-be19-d555288bd0cb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.391730] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cb05883-7c57-4fab-b215-c94ae1aaaf3e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.407810] env[62385]: DEBUG nova.compute.provider_tree [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 801.413420] env[62385]: DEBUG nova.network.neutron [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 801.579298] env[62385]: DEBUG oslo_concurrency.lockutils [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.580541] env[62385]: DEBUG nova.network.neutron [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Updating instance_info_cache with network_info: [{"id": "ff03b686-6596-418e-b748-0a8a6799a2b5", "address": "fa:16:3e:91:05:f2", "network": {"id": "afae6b07-d08c-413f-aa73-b545db1bb766", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1654965885-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "93111f5ba17c47de9691f90b99e03aa8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff03b686-65", "ovs_interfaceid": "ff03b686-6596-418e-b748-0a8a6799a2b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.606017] env[62385]: DEBUG oslo_vmware.api [None req-2fbab2d2-44d4-4ed4-9df6-51fefeaa7131 tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': task-1205754, 'name': PowerOffVM_Task, 'duration_secs': 0.191271} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.606853] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fbab2d2-44d4-4ed4-9df6-51fefeaa7131 tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 801.606957] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-2fbab2d2-44d4-4ed4-9df6-51fefeaa7131 tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 801.607187] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-db78b86d-2292-4ea4-8c72-3bbbf095a213 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.639106] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-2fbab2d2-44d4-4ed4-9df6-51fefeaa7131 tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 801.639509] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-2fbab2d2-44d4-4ed4-9df6-51fefeaa7131 tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Deleting contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 801.639567] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-2fbab2d2-44d4-4ed4-9df6-51fefeaa7131 tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Deleting the datastore file [datastore2] 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 801.639820] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3a806dc8-2507-427e-bc1b-e42e06d95f7f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.647424] env[62385]: DEBUG oslo_vmware.api [None req-2fbab2d2-44d4-4ed4-9df6-51fefeaa7131 tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Waiting for the task: (returnval){ [ 801.647424] env[62385]: value = "task-1205756" [ 801.647424] env[62385]: _type = "Task" [ 801.647424] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.656245] env[62385]: DEBUG oslo_vmware.api [None req-2fbab2d2-44d4-4ed4-9df6-51fefeaa7131 tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': task-1205756, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.913995] env[62385]: DEBUG nova.scheduler.client.report [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 801.925936] env[62385]: DEBUG nova.compute.manager [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] [instance: 2b744879-01d2-440e-8adf-58438b922888] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 801.957837] env[62385]: DEBUG nova.virt.hardware [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 801.958135] env[62385]: DEBUG nova.virt.hardware [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 801.958352] env[62385]: DEBUG nova.virt.hardware [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 801.958759] env[62385]: DEBUG nova.virt.hardware [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 801.958759] env[62385]: DEBUG nova.virt.hardware [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 801.958870] env[62385]: DEBUG nova.virt.hardware [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 801.959065] env[62385]: DEBUG nova.virt.hardware [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 801.959235] env[62385]: DEBUG nova.virt.hardware [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 801.959435] env[62385]: DEBUG nova.virt.hardware [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 801.959609] env[62385]: DEBUG nova.virt.hardware [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 801.959786] env[62385]: DEBUG nova.virt.hardware [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 801.960889] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09f40ec6-57f0-435e-bf24-912ce133454f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.970541] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd7985dc-ec3a-4b07-8c1f-c06f37e22268 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.083617] env[62385]: DEBUG oslo_concurrency.lockutils [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Releasing lock "refresh_cache-9d6f098a-0b05-43ef-96b0-9eb99ad3538c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.084279] env[62385]: DEBUG nova.compute.manager [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Instance network_info: |[{"id": "ff03b686-6596-418e-b748-0a8a6799a2b5", "address": "fa:16:3e:91:05:f2", "network": {"id": "afae6b07-d08c-413f-aa73-b545db1bb766", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1654965885-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "93111f5ba17c47de9691f90b99e03aa8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff03b686-65", "ovs_interfaceid": "ff03b686-6596-418e-b748-0a8a6799a2b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 802.084552] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:91:05:f2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4adc8ed0-d11a-4510-9be0-b27c0da3a903', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ff03b686-6596-418e-b748-0a8a6799a2b5', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 802.098026] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Creating folder: Project (93111f5ba17c47de9691f90b99e03aa8). Parent ref: group-v261107. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 802.098026] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3fe62ade-3d3b-424e-9710-4ce9c3f4c84a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.108238] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Created folder: Project (93111f5ba17c47de9691f90b99e03aa8) in parent group-v261107. [ 802.108554] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Creating folder: Instances. Parent ref: group-v261189. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 802.108822] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-88af0ee5-51d8-4303-b1a2-3a71f5a5937c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.121122] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Created folder: Instances in parent group-v261189. [ 802.121423] env[62385]: DEBUG oslo.service.loopingcall [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 802.121672] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 802.121914] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c97245ad-da9b-4740-900e-802ea32b7ff5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.146482] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 802.146482] env[62385]: value = "task-1205759" [ 802.146482] env[62385]: _type = "Task" [ 802.146482] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.166041] env[62385]: DEBUG oslo_vmware.api [None req-2fbab2d2-44d4-4ed4-9df6-51fefeaa7131 tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': task-1205756, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.170283] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205759, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.354152] env[62385]: DEBUG nova.compute.manager [req-cc2edd6f-b570-476f-845d-55b7541974c1 req-d1e9bca6-837a-41bd-bd9f-5812e440649e service nova] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Received event network-changed-ff03b686-6596-418e-b748-0a8a6799a2b5 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 802.354389] env[62385]: DEBUG nova.compute.manager [req-cc2edd6f-b570-476f-845d-55b7541974c1 req-d1e9bca6-837a-41bd-bd9f-5812e440649e service nova] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Refreshing instance network info cache due to event network-changed-ff03b686-6596-418e-b748-0a8a6799a2b5. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 802.354670] env[62385]: DEBUG oslo_concurrency.lockutils [req-cc2edd6f-b570-476f-845d-55b7541974c1 req-d1e9bca6-837a-41bd-bd9f-5812e440649e service nova] Acquiring lock "refresh_cache-9d6f098a-0b05-43ef-96b0-9eb99ad3538c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.354817] env[62385]: DEBUG oslo_concurrency.lockutils [req-cc2edd6f-b570-476f-845d-55b7541974c1 req-d1e9bca6-837a-41bd-bd9f-5812e440649e service nova] Acquired lock "refresh_cache-9d6f098a-0b05-43ef-96b0-9eb99ad3538c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.355029] env[62385]: DEBUG nova.network.neutron [req-cc2edd6f-b570-476f-845d-55b7541974c1 req-d1e9bca6-837a-41bd-bd9f-5812e440649e service nova] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Refreshing network info cache for port ff03b686-6596-418e-b748-0a8a6799a2b5 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 802.419022] env[62385]: DEBUG oslo_concurrency.lockutils [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.515s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.419626] env[62385]: DEBUG nova.compute.manager [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 802.423087] env[62385]: DEBUG oslo_concurrency.lockutils [None req-181b8245-1e74-4984-9c8b-2576aab7b28f tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.121s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.423349] env[62385]: DEBUG nova.objects.instance [None req-181b8245-1e74-4984-9c8b-2576aab7b28f tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Lazy-loading 'resources' on Instance uuid 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 802.660383] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205759, 'name': CreateVM_Task, 'duration_secs': 0.440897} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.661253] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 802.661990] env[62385]: DEBUG oslo_concurrency.lockutils [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.662184] env[62385]: DEBUG oslo_concurrency.lockutils [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.662514] env[62385]: DEBUG oslo_concurrency.lockutils [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 802.662775] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-466389a6-cdc3-40bd-8910-5da95f3c113e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.667603] env[62385]: DEBUG oslo_vmware.api [None req-2fbab2d2-44d4-4ed4-9df6-51fefeaa7131 tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Task: {'id': task-1205756, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.917339} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.668164] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-2fbab2d2-44d4-4ed4-9df6-51fefeaa7131 tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 802.668398] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-2fbab2d2-44d4-4ed4-9df6-51fefeaa7131 tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Deleted contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 802.668647] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-2fbab2d2-44d4-4ed4-9df6-51fefeaa7131 tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 802.668768] env[62385]: INFO nova.compute.manager [None req-2fbab2d2-44d4-4ed4-9df6-51fefeaa7131 tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Took 1.59 seconds to destroy the instance on the hypervisor. [ 802.669022] env[62385]: DEBUG oslo.service.loopingcall [None req-2fbab2d2-44d4-4ed4-9df6-51fefeaa7131 tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 802.669213] env[62385]: DEBUG nova.compute.manager [-] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 802.669324] env[62385]: DEBUG nova.network.neutron [-] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 802.671952] env[62385]: DEBUG oslo_vmware.api [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 802.671952] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52b7d2d0-2b89-0bc2-c46d-6b83f76a22e2" [ 802.671952] env[62385]: _type = "Task" [ 802.671952] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.680030] env[62385]: DEBUG oslo_vmware.api [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52b7d2d0-2b89-0bc2-c46d-6b83f76a22e2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.685810] env[62385]: DEBUG nova.network.neutron [-] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 802.927714] env[62385]: DEBUG nova.compute.utils [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 802.934918] env[62385]: DEBUG nova.compute.manager [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 802.935201] env[62385]: DEBUG nova.network.neutron [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 803.001664] env[62385]: DEBUG nova.policy [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd07da91a073f42a3886ee4b1dfda0c35', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f3248d9aaa7f4b7eb6d7b2bd01990bee', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 803.091660] env[62385]: DEBUG nova.network.neutron [req-cc2edd6f-b570-476f-845d-55b7541974c1 req-d1e9bca6-837a-41bd-bd9f-5812e440649e service nova] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Updated VIF entry in instance network info cache for port ff03b686-6596-418e-b748-0a8a6799a2b5. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 803.092028] env[62385]: DEBUG nova.network.neutron [req-cc2edd6f-b570-476f-845d-55b7541974c1 req-d1e9bca6-837a-41bd-bd9f-5812e440649e service nova] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Updating instance_info_cache with network_info: [{"id": "ff03b686-6596-418e-b748-0a8a6799a2b5", "address": "fa:16:3e:91:05:f2", "network": {"id": "afae6b07-d08c-413f-aa73-b545db1bb766", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1654965885-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "93111f5ba17c47de9691f90b99e03aa8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff03b686-65", "ovs_interfaceid": "ff03b686-6596-418e-b748-0a8a6799a2b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.190176] env[62385]: DEBUG nova.network.neutron [-] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.191575] env[62385]: DEBUG oslo_vmware.api [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52b7d2d0-2b89-0bc2-c46d-6b83f76a22e2, 'name': SearchDatastore_Task, 'duration_secs': 0.020392} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.191999] env[62385]: DEBUG oslo_concurrency.lockutils [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.193138] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 803.193696] env[62385]: DEBUG oslo_concurrency.lockutils [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.193931] env[62385]: DEBUG oslo_concurrency.lockutils [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.194157] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 803.194761] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dad143b9-62e8-416b-b8e3-7914973e19d6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.209491] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 803.209491] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 803.209899] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-25a4eed0-98bc-46dc-85bf-0762a250a3c8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.219706] env[62385]: DEBUG oslo_vmware.api [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 803.219706] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]527a9d3c-6ada-20f7-1c6b-02727f32e4db" [ 803.219706] env[62385]: _type = "Task" [ 803.219706] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.230993] env[62385]: DEBUG oslo_vmware.api [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]527a9d3c-6ada-20f7-1c6b-02727f32e4db, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.358019] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63b6e162-2cd7-4ae5-80d6-5ffbd2dd4640 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.365084] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-821ebd95-5b82-4060-ac6a-9aee53b20d23 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.399557] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ff03b82-f7e3-40ac-b3b7-d1c8723a29ab {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.408108] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a87f908a-70a7-4fe8-b6c9-03c92079f30d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.425709] env[62385]: DEBUG nova.compute.provider_tree [None req-181b8245-1e74-4984-9c8b-2576aab7b28f tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 803.433054] env[62385]: DEBUG nova.compute.manager [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 803.594682] env[62385]: DEBUG oslo_concurrency.lockutils [req-cc2edd6f-b570-476f-845d-55b7541974c1 req-d1e9bca6-837a-41bd-bd9f-5812e440649e service nova] Releasing lock "refresh_cache-9d6f098a-0b05-43ef-96b0-9eb99ad3538c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.595274] env[62385]: DEBUG nova.compute.manager [req-cc2edd6f-b570-476f-845d-55b7541974c1 req-d1e9bca6-837a-41bd-bd9f-5812e440649e service nova] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Received event network-changed-2c31d161-814f-4c23-9882-0aa436bb9e03 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 803.595274] env[62385]: DEBUG nova.compute.manager [req-cc2edd6f-b570-476f-845d-55b7541974c1 req-d1e9bca6-837a-41bd-bd9f-5812e440649e service nova] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Refreshing instance network info cache due to event network-changed-2c31d161-814f-4c23-9882-0aa436bb9e03. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 803.595393] env[62385]: DEBUG oslo_concurrency.lockutils [req-cc2edd6f-b570-476f-845d-55b7541974c1 req-d1e9bca6-837a-41bd-bd9f-5812e440649e service nova] Acquiring lock "refresh_cache-d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.595523] env[62385]: DEBUG oslo_concurrency.lockutils [req-cc2edd6f-b570-476f-845d-55b7541974c1 req-d1e9bca6-837a-41bd-bd9f-5812e440649e service nova] Acquired lock "refresh_cache-d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.595707] env[62385]: DEBUG nova.network.neutron [req-cc2edd6f-b570-476f-845d-55b7541974c1 req-d1e9bca6-837a-41bd-bd9f-5812e440649e service nova] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Refreshing network info cache for port 2c31d161-814f-4c23-9882-0aa436bb9e03 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 803.696055] env[62385]: INFO nova.compute.manager [-] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Took 1.03 seconds to deallocate network for instance. [ 803.734177] env[62385]: DEBUG oslo_vmware.api [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]527a9d3c-6ada-20f7-1c6b-02727f32e4db, 'name': SearchDatastore_Task, 'duration_secs': 0.012569} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.735077] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-926df0ed-8cb5-42d3-8e85-c744afb0be88 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.742704] env[62385]: DEBUG oslo_vmware.api [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 803.742704] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e4841c-591f-55a0-44c8-98e05010ea9f" [ 803.742704] env[62385]: _type = "Task" [ 803.742704] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.747385] env[62385]: DEBUG nova.network.neutron [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Successfully created port: 32c6bd5f-7ef8-454d-8360-1bc0fcb5b894 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 803.755213] env[62385]: DEBUG oslo_vmware.api [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e4841c-591f-55a0-44c8-98e05010ea9f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.840445] env[62385]: DEBUG nova.network.neutron [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] [instance: 2b744879-01d2-440e-8adf-58438b922888] Successfully updated port: f42dfa8b-5ef4-40f7-9359-aa425ae2fbad {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 803.928894] env[62385]: DEBUG nova.scheduler.client.report [None req-181b8245-1e74-4984-9c8b-2576aab7b28f tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 804.145645] env[62385]: DEBUG oslo_vmware.rw_handles [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5255f5df-ab9a-92a7-36ec-af386ab79ff8/disk-0.vmdk. {{(pid=62385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 804.146532] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be74db98-a1bb-4d60-8547-9bc6cacf8630 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.155335] env[62385]: DEBUG oslo_vmware.rw_handles [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5255f5df-ab9a-92a7-36ec-af386ab79ff8/disk-0.vmdk is in state: ready. {{(pid=62385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 804.155506] env[62385]: ERROR oslo_vmware.rw_handles [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5255f5df-ab9a-92a7-36ec-af386ab79ff8/disk-0.vmdk due to incomplete transfer. [ 804.155778] env[62385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-3dfc9dcf-1ae8-4277-8865-c5a68c04add0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.164150] env[62385]: DEBUG oslo_vmware.rw_handles [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5255f5df-ab9a-92a7-36ec-af386ab79ff8/disk-0.vmdk. {{(pid=62385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 804.165158] env[62385]: DEBUG nova.virt.vmwareapi.images [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Uploaded image 28813632-55e0-4297-8acd-1f1992ee8fe5 to the Glance image server {{(pid=62385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 804.167372] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Destroying the VM {{(pid=62385) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 804.167639] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-30b626f8-d40d-4bd9-9ce0-5f4d86053c77 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.174850] env[62385]: DEBUG oslo_vmware.api [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Waiting for the task: (returnval){ [ 804.174850] env[62385]: value = "task-1205760" [ 804.174850] env[62385]: _type = "Task" [ 804.174850] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.185508] env[62385]: DEBUG oslo_vmware.api [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205760, 'name': Destroy_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.203637] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2fbab2d2-44d4-4ed4-9df6-51fefeaa7131 tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.255200] env[62385]: DEBUG oslo_vmware.api [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e4841c-591f-55a0-44c8-98e05010ea9f, 'name': SearchDatastore_Task, 'duration_secs': 0.028811} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.255481] env[62385]: DEBUG oslo_concurrency.lockutils [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.256510] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 9d6f098a-0b05-43ef-96b0-9eb99ad3538c/9d6f098a-0b05-43ef-96b0-9eb99ad3538c.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 804.256510] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8800a75b-f5c8-460e-bbab-32d5efead19c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.266290] env[62385]: DEBUG oslo_vmware.api [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 804.266290] env[62385]: value = "task-1205761" [ 804.266290] env[62385]: _type = "Task" [ 804.266290] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.274395] env[62385]: DEBUG oslo_vmware.api [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1205761, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.343975] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Acquiring lock "refresh_cache-2b744879-01d2-440e-8adf-58438b922888" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.344113] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Acquired lock "refresh_cache-2b744879-01d2-440e-8adf-58438b922888" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.344268] env[62385]: DEBUG nova.network.neutron [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] [instance: 2b744879-01d2-440e-8adf-58438b922888] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 804.389606] env[62385]: DEBUG nova.compute.manager [req-0bd0b3bf-6ade-498c-a4f3-9826b1e2ff36 req-f006805c-a71b-42ba-b0c8-10c514c6501c service nova] [instance: 2b744879-01d2-440e-8adf-58438b922888] Received event network-vif-plugged-f42dfa8b-5ef4-40f7-9359-aa425ae2fbad {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 804.389847] env[62385]: DEBUG oslo_concurrency.lockutils [req-0bd0b3bf-6ade-498c-a4f3-9826b1e2ff36 req-f006805c-a71b-42ba-b0c8-10c514c6501c service nova] Acquiring lock "2b744879-01d2-440e-8adf-58438b922888-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.390205] env[62385]: DEBUG oslo_concurrency.lockutils [req-0bd0b3bf-6ade-498c-a4f3-9826b1e2ff36 req-f006805c-a71b-42ba-b0c8-10c514c6501c service nova] Lock "2b744879-01d2-440e-8adf-58438b922888-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.390361] env[62385]: DEBUG oslo_concurrency.lockutils [req-0bd0b3bf-6ade-498c-a4f3-9826b1e2ff36 req-f006805c-a71b-42ba-b0c8-10c514c6501c service nova] Lock "2b744879-01d2-440e-8adf-58438b922888-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.390579] env[62385]: DEBUG nova.compute.manager [req-0bd0b3bf-6ade-498c-a4f3-9826b1e2ff36 req-f006805c-a71b-42ba-b0c8-10c514c6501c service nova] [instance: 2b744879-01d2-440e-8adf-58438b922888] No waiting events found dispatching network-vif-plugged-f42dfa8b-5ef4-40f7-9359-aa425ae2fbad {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 804.390818] env[62385]: WARNING nova.compute.manager [req-0bd0b3bf-6ade-498c-a4f3-9826b1e2ff36 req-f006805c-a71b-42ba-b0c8-10c514c6501c service nova] [instance: 2b744879-01d2-440e-8adf-58438b922888] Received unexpected event network-vif-plugged-f42dfa8b-5ef4-40f7-9359-aa425ae2fbad for instance with vm_state building and task_state spawning. [ 804.390987] env[62385]: DEBUG nova.compute.manager [req-0bd0b3bf-6ade-498c-a4f3-9826b1e2ff36 req-f006805c-a71b-42ba-b0c8-10c514c6501c service nova] [instance: 2b744879-01d2-440e-8adf-58438b922888] Received event network-changed-f42dfa8b-5ef4-40f7-9359-aa425ae2fbad {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 804.391188] env[62385]: DEBUG nova.compute.manager [req-0bd0b3bf-6ade-498c-a4f3-9826b1e2ff36 req-f006805c-a71b-42ba-b0c8-10c514c6501c service nova] [instance: 2b744879-01d2-440e-8adf-58438b922888] Refreshing instance network info cache due to event network-changed-f42dfa8b-5ef4-40f7-9359-aa425ae2fbad. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 804.391397] env[62385]: DEBUG oslo_concurrency.lockutils [req-0bd0b3bf-6ade-498c-a4f3-9826b1e2ff36 req-f006805c-a71b-42ba-b0c8-10c514c6501c service nova] Acquiring lock "refresh_cache-2b744879-01d2-440e-8adf-58438b922888" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.405829] env[62385]: DEBUG nova.network.neutron [req-cc2edd6f-b570-476f-845d-55b7541974c1 req-d1e9bca6-837a-41bd-bd9f-5812e440649e service nova] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Updated VIF entry in instance network info cache for port 2c31d161-814f-4c23-9882-0aa436bb9e03. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 804.406192] env[62385]: DEBUG nova.network.neutron [req-cc2edd6f-b570-476f-845d-55b7541974c1 req-d1e9bca6-837a-41bd-bd9f-5812e440649e service nova] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Updating instance_info_cache with network_info: [{"id": "2c31d161-814f-4c23-9882-0aa436bb9e03", "address": "fa:16:3e:d7:43:42", "network": {"id": "aca904e1-fbd6-4def-bbec-bd70c0fc8af9", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-249813742-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.141", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e99f1e4cfd84bf083c8261b2ffcfa2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f7abe5d-b8fe-4983-bd50-e7469f1fe7f3", "external-id": "nsx-vlan-transportzone-263", "segmentation_id": 263, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c31d161-81", "ovs_interfaceid": "2c31d161-814f-4c23-9882-0aa436bb9e03", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.433772] env[62385]: DEBUG oslo_concurrency.lockutils [None req-181b8245-1e74-4984-9c8b-2576aab7b28f tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.011s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.436792] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.899s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.438115] env[62385]: INFO nova.compute.claims [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 804.445170] env[62385]: DEBUG nova.compute.manager [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 804.467223] env[62385]: INFO nova.scheduler.client.report [None req-181b8245-1e74-4984-9c8b-2576aab7b28f tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Deleted allocations for instance 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69 [ 804.476900] env[62385]: DEBUG nova.virt.hardware [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 804.477312] env[62385]: DEBUG nova.virt.hardware [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 804.477576] env[62385]: DEBUG nova.virt.hardware [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 804.477906] env[62385]: DEBUG nova.virt.hardware [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 804.478178] env[62385]: DEBUG nova.virt.hardware [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 804.478624] env[62385]: DEBUG nova.virt.hardware [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 804.478820] env[62385]: DEBUG nova.virt.hardware [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 804.479126] env[62385]: DEBUG nova.virt.hardware [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 804.479448] env[62385]: DEBUG nova.virt.hardware [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 804.479757] env[62385]: DEBUG nova.virt.hardware [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 804.480077] env[62385]: DEBUG nova.virt.hardware [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 804.481422] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2e23757-3600-4662-8108-1e3d07060256 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.513350] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f99d6125-7d06-4413-8b7d-0c0f7c22aaa2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.687427] env[62385]: DEBUG oslo_vmware.api [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205760, 'name': Destroy_Task, 'duration_secs': 0.335641} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.687701] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Destroyed the VM [ 804.688017] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Deleting Snapshot of the VM instance {{(pid=62385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 804.688344] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-d87f8e26-23d7-408f-9dba-3f3b4b82eeef {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.697875] env[62385]: DEBUG oslo_vmware.api [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Waiting for the task: (returnval){ [ 804.697875] env[62385]: value = "task-1205762" [ 804.697875] env[62385]: _type = "Task" [ 804.697875] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.708854] env[62385]: DEBUG oslo_vmware.api [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205762, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.781265] env[62385]: DEBUG oslo_vmware.api [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1205761, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.896585] env[62385]: DEBUG nova.network.neutron [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] [instance: 2b744879-01d2-440e-8adf-58438b922888] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 804.908740] env[62385]: DEBUG oslo_concurrency.lockutils [req-cc2edd6f-b570-476f-845d-55b7541974c1 req-d1e9bca6-837a-41bd-bd9f-5812e440649e service nova] Releasing lock "refresh_cache-d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.975713] env[62385]: DEBUG oslo_concurrency.lockutils [None req-181b8245-1e74-4984-9c8b-2576aab7b28f tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Lock "4a4b254c-46a3-4db6-9c0f-f476ae8f2b69" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.687s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.124844] env[62385]: DEBUG nova.network.neutron [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] [instance: 2b744879-01d2-440e-8adf-58438b922888] Updating instance_info_cache with network_info: [{"id": "f42dfa8b-5ef4-40f7-9359-aa425ae2fbad", "address": "fa:16:3e:13:3a:d3", "network": {"id": "4626c8bb-88ee-451d-beba-6455174716bf", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-2117107750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "073bfa66c3924f9db4f43ce2efff97b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2c2daf7c-c01b-41b1-a09a-fb8b893b4c80", "external-id": "nsx-vlan-transportzone-89", "segmentation_id": 89, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf42dfa8b-5e", "ovs_interfaceid": "f42dfa8b-5ef4-40f7-9359-aa425ae2fbad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.139501] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5d6a954b-460d-4b98-b7c6-e84c4438007a tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Acquiring lock "1ddb4d6d-3872-49db-bb40-e21721241e89" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.139801] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5d6a954b-460d-4b98-b7c6-e84c4438007a tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Lock "1ddb4d6d-3872-49db-bb40-e21721241e89" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.140162] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5d6a954b-460d-4b98-b7c6-e84c4438007a tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Acquiring lock "1ddb4d6d-3872-49db-bb40-e21721241e89-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.140402] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5d6a954b-460d-4b98-b7c6-e84c4438007a tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Lock "1ddb4d6d-3872-49db-bb40-e21721241e89-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.140578] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5d6a954b-460d-4b98-b7c6-e84c4438007a tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Lock "1ddb4d6d-3872-49db-bb40-e21721241e89-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.142763] env[62385]: INFO nova.compute.manager [None req-5d6a954b-460d-4b98-b7c6-e84c4438007a tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Terminating instance [ 805.147088] env[62385]: DEBUG nova.compute.manager [None req-5d6a954b-460d-4b98-b7c6-e84c4438007a tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 805.147298] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-5d6a954b-460d-4b98-b7c6-e84c4438007a tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 805.148459] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1856517-6d29-41dd-9d77-66c4ec72e449 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.157540] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d6a954b-460d-4b98-b7c6-e84c4438007a tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 805.157882] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c632200f-39de-4cb9-b7cb-a7933c749866 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.165202] env[62385]: DEBUG oslo_vmware.api [None req-5d6a954b-460d-4b98-b7c6-e84c4438007a tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Waiting for the task: (returnval){ [ 805.165202] env[62385]: value = "task-1205763" [ 805.165202] env[62385]: _type = "Task" [ 805.165202] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.174910] env[62385]: DEBUG oslo_vmware.api [None req-5d6a954b-460d-4b98-b7c6-e84c4438007a tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205763, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.208067] env[62385]: DEBUG oslo_vmware.api [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205762, 'name': RemoveSnapshot_Task} progress is 15%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.280221] env[62385]: DEBUG oslo_vmware.api [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1205761, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.555596} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.280574] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 9d6f098a-0b05-43ef-96b0-9eb99ad3538c/9d6f098a-0b05-43ef-96b0-9eb99ad3538c.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 805.280853] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 805.281145] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9cbb03cb-3dfc-4c3f-925b-5bd359a73ff2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.288832] env[62385]: DEBUG oslo_vmware.api [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 805.288832] env[62385]: value = "task-1205764" [ 805.288832] env[62385]: _type = "Task" [ 805.288832] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.299356] env[62385]: DEBUG oslo_vmware.api [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1205764, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.627810] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Releasing lock "refresh_cache-2b744879-01d2-440e-8adf-58438b922888" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.628183] env[62385]: DEBUG nova.compute.manager [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] [instance: 2b744879-01d2-440e-8adf-58438b922888] Instance network_info: |[{"id": "f42dfa8b-5ef4-40f7-9359-aa425ae2fbad", "address": "fa:16:3e:13:3a:d3", "network": {"id": "4626c8bb-88ee-451d-beba-6455174716bf", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-2117107750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "073bfa66c3924f9db4f43ce2efff97b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2c2daf7c-c01b-41b1-a09a-fb8b893b4c80", "external-id": "nsx-vlan-transportzone-89", "segmentation_id": 89, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf42dfa8b-5e", "ovs_interfaceid": "f42dfa8b-5ef4-40f7-9359-aa425ae2fbad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 805.628565] env[62385]: DEBUG oslo_concurrency.lockutils [req-0bd0b3bf-6ade-498c-a4f3-9826b1e2ff36 req-f006805c-a71b-42ba-b0c8-10c514c6501c service nova] Acquired lock "refresh_cache-2b744879-01d2-440e-8adf-58438b922888" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.628755] env[62385]: DEBUG nova.network.neutron [req-0bd0b3bf-6ade-498c-a4f3-9826b1e2ff36 req-f006805c-a71b-42ba-b0c8-10c514c6501c service nova] [instance: 2b744879-01d2-440e-8adf-58438b922888] Refreshing network info cache for port f42dfa8b-5ef4-40f7-9359-aa425ae2fbad {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 805.629973] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] [instance: 2b744879-01d2-440e-8adf-58438b922888] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:13:3a:d3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2c2daf7c-c01b-41b1-a09a-fb8b893b4c80', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f42dfa8b-5ef4-40f7-9359-aa425ae2fbad', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 805.637473] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Creating folder: Project (073bfa66c3924f9db4f43ce2efff97b2). Parent ref: group-v261107. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 805.643113] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c4bc0526-517c-45ac-b242-6f6feefa1c7b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.656918] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Created folder: Project (073bfa66c3924f9db4f43ce2efff97b2) in parent group-v261107. [ 805.657169] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Creating folder: Instances. Parent ref: group-v261192. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 805.657435] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1b4e09d2-6a44-415d-bcec-657eb9efe389 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.674719] env[62385]: DEBUG oslo_vmware.api [None req-5d6a954b-460d-4b98-b7c6-e84c4438007a tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205763, 'name': PowerOffVM_Task, 'duration_secs': 0.363604} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.678640] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d6a954b-460d-4b98-b7c6-e84c4438007a tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 805.678824] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-5d6a954b-460d-4b98-b7c6-e84c4438007a tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 805.679116] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Created folder: Instances in parent group-v261192. [ 805.679340] env[62385]: DEBUG oslo.service.loopingcall [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 805.679803] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d102d1f2-6d3b-4ab3-ad33-8591ecd769c0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.681481] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2b744879-01d2-440e-8adf-58438b922888] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 805.681923] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7b7efee6-44cf-447e-93f3-47d43e093fef {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.710782] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 805.710782] env[62385]: value = "task-1205768" [ 805.710782] env[62385]: _type = "Task" [ 805.710782] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.711129] env[62385]: DEBUG oslo_vmware.api [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205762, 'name': RemoveSnapshot_Task} progress is 70%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.722224] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205768, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.784842] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-5d6a954b-460d-4b98-b7c6-e84c4438007a tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 805.785213] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-5d6a954b-460d-4b98-b7c6-e84c4438007a tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 805.785484] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d6a954b-460d-4b98-b7c6-e84c4438007a tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Deleting the datastore file [datastore1] 1ddb4d6d-3872-49db-bb40-e21721241e89 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 805.785796] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-74c440ae-4750-4905-8325-a3c59e8819e1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.796234] env[62385]: DEBUG oslo_vmware.api [None req-5d6a954b-460d-4b98-b7c6-e84c4438007a tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Waiting for the task: (returnval){ [ 805.796234] env[62385]: value = "task-1205769" [ 805.796234] env[62385]: _type = "Task" [ 805.796234] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.802905] env[62385]: DEBUG oslo_vmware.api [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1205764, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.262466} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.803603] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 805.804499] env[62385]: DEBUG nova.network.neutron [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Successfully updated port: 32c6bd5f-7ef8-454d-8360-1bc0fcb5b894 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 805.809635] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-749d0089-87f4-4024-9b1a-036da85b890a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.813716] env[62385]: DEBUG oslo_vmware.api [None req-5d6a954b-460d-4b98-b7c6-e84c4438007a tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205769, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.840512] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] 9d6f098a-0b05-43ef-96b0-9eb99ad3538c/9d6f098a-0b05-43ef-96b0-9eb99ad3538c.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 805.845658] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b9aa99b1-0664-45f8-a564-92dcfb7c4cec {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.866980] env[62385]: DEBUG oslo_vmware.api [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 805.866980] env[62385]: value = "task-1205770" [ 805.866980] env[62385]: _type = "Task" [ 805.866980] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.878841] env[62385]: DEBUG oslo_vmware.api [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1205770, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.915174] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35d0c38e-40a5-4221-b9be-084654152e21 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.923952] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1a6a7ca-8172-4cdf-8c55-e609cb936cbc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.960028] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce613f21-043c-44e1-9b3d-267c0184df12 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.969905] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65a4091a-9578-46e1-8ebf-ac679d0f4a25 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.986146] env[62385]: DEBUG nova.compute.provider_tree [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 806.128398] env[62385]: DEBUG nova.network.neutron [req-0bd0b3bf-6ade-498c-a4f3-9826b1e2ff36 req-f006805c-a71b-42ba-b0c8-10c514c6501c service nova] [instance: 2b744879-01d2-440e-8adf-58438b922888] Updated VIF entry in instance network info cache for port f42dfa8b-5ef4-40f7-9359-aa425ae2fbad. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 806.128825] env[62385]: DEBUG nova.network.neutron [req-0bd0b3bf-6ade-498c-a4f3-9826b1e2ff36 req-f006805c-a71b-42ba-b0c8-10c514c6501c service nova] [instance: 2b744879-01d2-440e-8adf-58438b922888] Updating instance_info_cache with network_info: [{"id": "f42dfa8b-5ef4-40f7-9359-aa425ae2fbad", "address": "fa:16:3e:13:3a:d3", "network": {"id": "4626c8bb-88ee-451d-beba-6455174716bf", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-2117107750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "073bfa66c3924f9db4f43ce2efff97b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2c2daf7c-c01b-41b1-a09a-fb8b893b4c80", "external-id": "nsx-vlan-transportzone-89", "segmentation_id": 89, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf42dfa8b-5e", "ovs_interfaceid": "f42dfa8b-5ef4-40f7-9359-aa425ae2fbad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.210616] env[62385]: DEBUG oslo_vmware.api [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205762, 'name': RemoveSnapshot_Task, 'duration_secs': 1.387276} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.210989] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Deleted Snapshot of the VM instance {{(pid=62385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 806.211254] env[62385]: INFO nova.compute.manager [None req-f007568d-6057-406a-97bd-7332e27a25aa tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Took 15.17 seconds to snapshot the instance on the hypervisor. [ 806.225036] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205768, 'name': CreateVM_Task, 'duration_secs': 0.407805} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.225036] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2b744879-01d2-440e-8adf-58438b922888] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 806.225036] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.225036] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.225036] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 806.225426] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-95990ed5-0c34-44c4-99d8-f7db0c673642 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.231582] env[62385]: DEBUG oslo_vmware.api [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Waiting for the task: (returnval){ [ 806.231582] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e2adb6-f4fb-64cb-2caa-c4128c9bdae9" [ 806.231582] env[62385]: _type = "Task" [ 806.231582] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.240519] env[62385]: DEBUG oslo_vmware.api [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e2adb6-f4fb-64cb-2caa-c4128c9bdae9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.307348] env[62385]: DEBUG oslo_vmware.api [None req-5d6a954b-460d-4b98-b7c6-e84c4438007a tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Task: {'id': task-1205769, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.198882} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.307468] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d6a954b-460d-4b98-b7c6-e84c4438007a tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 806.307655] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-5d6a954b-460d-4b98-b7c6-e84c4438007a tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 806.307836] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-5d6a954b-460d-4b98-b7c6-e84c4438007a tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 806.308024] env[62385]: INFO nova.compute.manager [None req-5d6a954b-460d-4b98-b7c6-e84c4438007a tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Took 1.16 seconds to destroy the instance on the hypervisor. [ 806.308276] env[62385]: DEBUG oslo.service.loopingcall [None req-5d6a954b-460d-4b98-b7c6-e84c4438007a tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 806.308525] env[62385]: DEBUG nova.compute.manager [-] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 806.308633] env[62385]: DEBUG nova.network.neutron [-] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 806.313986] env[62385]: DEBUG oslo_concurrency.lockutils [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Acquiring lock "refresh_cache-e4089e71-eea1-4e16-b90c-97966333f2bc" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.314150] env[62385]: DEBUG oslo_concurrency.lockutils [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Acquired lock "refresh_cache-e4089e71-eea1-4e16-b90c-97966333f2bc" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.314268] env[62385]: DEBUG nova.network.neutron [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 806.379993] env[62385]: DEBUG oslo_vmware.api [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1205770, 'name': ReconfigVM_Task, 'duration_secs': 0.351898} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.380374] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Reconfigured VM instance instance-0000003f to attach disk [datastore2] 9d6f098a-0b05-43ef-96b0-9eb99ad3538c/9d6f098a-0b05-43ef-96b0-9eb99ad3538c.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 806.381239] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6a91d35b-cc99-4a3e-b648-832876541a10 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.391628] env[62385]: DEBUG oslo_vmware.api [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 806.391628] env[62385]: value = "task-1205771" [ 806.391628] env[62385]: _type = "Task" [ 806.391628] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.403368] env[62385]: DEBUG oslo_vmware.api [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1205771, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.419914] env[62385]: DEBUG nova.compute.manager [req-d234aa25-e83a-4fbb-9eba-45c5ccd6117e req-c1729eb2-101d-4a79-a4b4-674944797b37 service nova] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Received event network-vif-plugged-32c6bd5f-7ef8-454d-8360-1bc0fcb5b894 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 806.420124] env[62385]: DEBUG oslo_concurrency.lockutils [req-d234aa25-e83a-4fbb-9eba-45c5ccd6117e req-c1729eb2-101d-4a79-a4b4-674944797b37 service nova] Acquiring lock "e4089e71-eea1-4e16-b90c-97966333f2bc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.420499] env[62385]: DEBUG oslo_concurrency.lockutils [req-d234aa25-e83a-4fbb-9eba-45c5ccd6117e req-c1729eb2-101d-4a79-a4b4-674944797b37 service nova] Lock "e4089e71-eea1-4e16-b90c-97966333f2bc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.420596] env[62385]: DEBUG oslo_concurrency.lockutils [req-d234aa25-e83a-4fbb-9eba-45c5ccd6117e req-c1729eb2-101d-4a79-a4b4-674944797b37 service nova] Lock "e4089e71-eea1-4e16-b90c-97966333f2bc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.420721] env[62385]: DEBUG nova.compute.manager [req-d234aa25-e83a-4fbb-9eba-45c5ccd6117e req-c1729eb2-101d-4a79-a4b4-674944797b37 service nova] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] No waiting events found dispatching network-vif-plugged-32c6bd5f-7ef8-454d-8360-1bc0fcb5b894 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 806.420885] env[62385]: WARNING nova.compute.manager [req-d234aa25-e83a-4fbb-9eba-45c5ccd6117e req-c1729eb2-101d-4a79-a4b4-674944797b37 service nova] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Received unexpected event network-vif-plugged-32c6bd5f-7ef8-454d-8360-1bc0fcb5b894 for instance with vm_state building and task_state spawning. [ 806.421104] env[62385]: DEBUG nova.compute.manager [req-d234aa25-e83a-4fbb-9eba-45c5ccd6117e req-c1729eb2-101d-4a79-a4b4-674944797b37 service nova] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Received event network-changed-32c6bd5f-7ef8-454d-8360-1bc0fcb5b894 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 806.421229] env[62385]: DEBUG nova.compute.manager [req-d234aa25-e83a-4fbb-9eba-45c5ccd6117e req-c1729eb2-101d-4a79-a4b4-674944797b37 service nova] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Refreshing instance network info cache due to event network-changed-32c6bd5f-7ef8-454d-8360-1bc0fcb5b894. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 806.421404] env[62385]: DEBUG oslo_concurrency.lockutils [req-d234aa25-e83a-4fbb-9eba-45c5ccd6117e req-c1729eb2-101d-4a79-a4b4-674944797b37 service nova] Acquiring lock "refresh_cache-e4089e71-eea1-4e16-b90c-97966333f2bc" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.489419] env[62385]: DEBUG nova.scheduler.client.report [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 806.630936] env[62385]: DEBUG oslo_concurrency.lockutils [req-0bd0b3bf-6ade-498c-a4f3-9826b1e2ff36 req-f006805c-a71b-42ba-b0c8-10c514c6501c service nova] Releasing lock "refresh_cache-2b744879-01d2-440e-8adf-58438b922888" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.743345] env[62385]: DEBUG oslo_vmware.api [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e2adb6-f4fb-64cb-2caa-c4128c9bdae9, 'name': SearchDatastore_Task, 'duration_secs': 0.010552} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.743657] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.743894] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] [instance: 2b744879-01d2-440e-8adf-58438b922888] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 806.744153] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.744308] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.744493] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 806.744870] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-54e9363d-5f66-4e3f-b2cf-b0e39c104d0f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.754694] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 806.754893] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 806.756089] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb07b3a7-fd40-4028-b1d0-6b59f5b768dd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.761920] env[62385]: DEBUG oslo_vmware.api [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Waiting for the task: (returnval){ [ 806.761920] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52aee438-fbba-f671-f0a7-bd6870bed1d7" [ 806.761920] env[62385]: _type = "Task" [ 806.761920] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.770388] env[62385]: DEBUG oslo_vmware.api [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52aee438-fbba-f671-f0a7-bd6870bed1d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.848116] env[62385]: DEBUG nova.network.neutron [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 806.904190] env[62385]: DEBUG oslo_vmware.api [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1205771, 'name': Rename_Task, 'duration_secs': 0.155698} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.904502] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 806.904762] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8f888ddc-4adc-4972-8cee-1accb958d8b2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.912460] env[62385]: DEBUG oslo_vmware.api [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 806.912460] env[62385]: value = "task-1205772" [ 806.912460] env[62385]: _type = "Task" [ 806.912460] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.921429] env[62385]: DEBUG oslo_vmware.api [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1205772, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.994703] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.559s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.995268] env[62385]: DEBUG nova.compute.manager [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 806.997865] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.644s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.999356] env[62385]: INFO nova.compute.claims [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 807.119776] env[62385]: DEBUG nova.network.neutron [-] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.274015] env[62385]: DEBUG oslo_vmware.api [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52aee438-fbba-f671-f0a7-bd6870bed1d7, 'name': SearchDatastore_Task, 'duration_secs': 0.010457} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.274752] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-873a605b-7f93-4f30-b5cb-5b616d9f93c2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.281529] env[62385]: DEBUG oslo_vmware.api [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Waiting for the task: (returnval){ [ 807.281529] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52d1af46-88be-846b-ae7e-7f556456ecaa" [ 807.281529] env[62385]: _type = "Task" [ 807.281529] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.292691] env[62385]: DEBUG oslo_vmware.api [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52d1af46-88be-846b-ae7e-7f556456ecaa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.331059] env[62385]: DEBUG nova.network.neutron [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Updating instance_info_cache with network_info: [{"id": "32c6bd5f-7ef8-454d-8360-1bc0fcb5b894", "address": "fa:16:3e:82:b0:a8", "network": {"id": "5b5fe53c-e8fb-412e-b86d-a102146875c0", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.208", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "2e3b444561ba4b12b2f10b6744f6aa39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d5970ab5-34b8-4065-bfa6-f568b8f103b7", "external-id": "nsx-vlan-transportzone-418", "segmentation_id": 418, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32c6bd5f-7e", "ovs_interfaceid": "32c6bd5f-7ef8-454d-8360-1bc0fcb5b894", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.424454] env[62385]: DEBUG oslo_vmware.api [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1205772, 'name': PowerOnVM_Task, 'duration_secs': 0.47163} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.424688] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 807.424914] env[62385]: INFO nova.compute.manager [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Took 8.01 seconds to spawn the instance on the hypervisor. [ 807.425128] env[62385]: DEBUG nova.compute.manager [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 807.425911] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6daff899-fbab-48a6-b9e4-a18258feefca {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.442268] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4e6ebd10-fb45-4b48-ad7e-266e6c10e531 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Acquiring lock "a5c4afc6-38a6-4815-8ec4-cc01c24489bf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.442514] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4e6ebd10-fb45-4b48-ad7e-266e6c10e531 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Lock "a5c4afc6-38a6-4815-8ec4-cc01c24489bf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.442976] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4e6ebd10-fb45-4b48-ad7e-266e6c10e531 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Acquiring lock "a5c4afc6-38a6-4815-8ec4-cc01c24489bf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.443210] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4e6ebd10-fb45-4b48-ad7e-266e6c10e531 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Lock "a5c4afc6-38a6-4815-8ec4-cc01c24489bf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.444585] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4e6ebd10-fb45-4b48-ad7e-266e6c10e531 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Lock "a5c4afc6-38a6-4815-8ec4-cc01c24489bf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.445475] env[62385]: INFO nova.compute.manager [None req-4e6ebd10-fb45-4b48-ad7e-266e6c10e531 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Terminating instance [ 807.447404] env[62385]: DEBUG nova.compute.manager [None req-4e6ebd10-fb45-4b48-ad7e-266e6c10e531 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 807.447603] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4e6ebd10-fb45-4b48-ad7e-266e6c10e531 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 807.448524] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6c98926-9690-41ce-8bfd-d80730db4e18 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.456767] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e6ebd10-fb45-4b48-ad7e-266e6c10e531 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 807.457067] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b67e0ec0-154b-42c6-978e-41aa819e040f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.466510] env[62385]: DEBUG oslo_vmware.api [None req-4e6ebd10-fb45-4b48-ad7e-266e6c10e531 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Waiting for the task: (returnval){ [ 807.466510] env[62385]: value = "task-1205773" [ 807.466510] env[62385]: _type = "Task" [ 807.466510] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.476169] env[62385]: DEBUG oslo_vmware.api [None req-4e6ebd10-fb45-4b48-ad7e-266e6c10e531 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205773, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.505825] env[62385]: DEBUG nova.compute.utils [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 807.510184] env[62385]: DEBUG nova.compute.manager [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 807.510184] env[62385]: DEBUG nova.network.neutron [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 807.592338] env[62385]: DEBUG nova.policy [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd07da91a073f42a3886ee4b1dfda0c35', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f3248d9aaa7f4b7eb6d7b2bd01990bee', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 807.622759] env[62385]: INFO nova.compute.manager [-] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Took 1.31 seconds to deallocate network for instance. [ 807.793638] env[62385]: DEBUG oslo_vmware.api [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52d1af46-88be-846b-ae7e-7f556456ecaa, 'name': SearchDatastore_Task, 'duration_secs': 0.017495} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.793638] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.793967] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 2b744879-01d2-440e-8adf-58438b922888/2b744879-01d2-440e-8adf-58438b922888.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 807.794085] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0b98287a-247d-46ae-934b-d5cacae2770d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.802829] env[62385]: DEBUG oslo_vmware.api [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Waiting for the task: (returnval){ [ 807.802829] env[62385]: value = "task-1205774" [ 807.802829] env[62385]: _type = "Task" [ 807.802829] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.812573] env[62385]: DEBUG oslo_vmware.api [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Task: {'id': task-1205774, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.834386] env[62385]: DEBUG oslo_concurrency.lockutils [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Releasing lock "refresh_cache-e4089e71-eea1-4e16-b90c-97966333f2bc" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.834817] env[62385]: DEBUG nova.compute.manager [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Instance network_info: |[{"id": "32c6bd5f-7ef8-454d-8360-1bc0fcb5b894", "address": "fa:16:3e:82:b0:a8", "network": {"id": "5b5fe53c-e8fb-412e-b86d-a102146875c0", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.208", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "2e3b444561ba4b12b2f10b6744f6aa39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d5970ab5-34b8-4065-bfa6-f568b8f103b7", "external-id": "nsx-vlan-transportzone-418", "segmentation_id": 418, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32c6bd5f-7e", "ovs_interfaceid": "32c6bd5f-7ef8-454d-8360-1bc0fcb5b894", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 807.835200] env[62385]: DEBUG oslo_concurrency.lockutils [req-d234aa25-e83a-4fbb-9eba-45c5ccd6117e req-c1729eb2-101d-4a79-a4b4-674944797b37 service nova] Acquired lock "refresh_cache-e4089e71-eea1-4e16-b90c-97966333f2bc" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.835450] env[62385]: DEBUG nova.network.neutron [req-d234aa25-e83a-4fbb-9eba-45c5ccd6117e req-c1729eb2-101d-4a79-a4b4-674944797b37 service nova] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Refreshing network info cache for port 32c6bd5f-7ef8-454d-8360-1bc0fcb5b894 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 807.836809] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:82:b0:a8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd5970ab5-34b8-4065-bfa6-f568b8f103b7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '32c6bd5f-7ef8-454d-8360-1bc0fcb5b894', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 807.852945] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Creating folder: Project (f3248d9aaa7f4b7eb6d7b2bd01990bee). Parent ref: group-v261107. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 807.855055] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9f3df6ff-051b-4f25-8d06-8d63b94af405 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.868388] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Created folder: Project (f3248d9aaa7f4b7eb6d7b2bd01990bee) in parent group-v261107. [ 807.868588] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Creating folder: Instances. Parent ref: group-v261195. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 807.868842] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4dcfde3b-ee05-42f3-afb7-249f24f676ce {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.879257] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Created folder: Instances in parent group-v261195. [ 807.879522] env[62385]: DEBUG oslo.service.loopingcall [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 807.879723] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 807.879944] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c938b4ef-cd3d-4891-b98d-0278a792c774 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.901825] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 807.901825] env[62385]: value = "task-1205777" [ 807.901825] env[62385]: _type = "Task" [ 807.901825] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.911513] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205777, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.943728] env[62385]: INFO nova.compute.manager [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Took 40.64 seconds to build instance. [ 807.976204] env[62385]: DEBUG oslo_vmware.api [None req-4e6ebd10-fb45-4b48-ad7e-266e6c10e531 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205773, 'name': PowerOffVM_Task, 'duration_secs': 0.419653} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.976469] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e6ebd10-fb45-4b48-ad7e-266e6c10e531 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 807.976796] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4e6ebd10-fb45-4b48-ad7e-266e6c10e531 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 807.976796] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-05bc8ae2-5a07-45a3-8ec3-444deeffc5b6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.015020] env[62385]: DEBUG nova.compute.manager [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 808.042392] env[62385]: DEBUG nova.network.neutron [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Successfully created port: 936e490f-ca20-418c-9b14-7e99db5ee2a9 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 808.052929] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4e6ebd10-fb45-4b48-ad7e-266e6c10e531 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 808.056203] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4e6ebd10-fb45-4b48-ad7e-266e6c10e531 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 808.056203] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e6ebd10-fb45-4b48-ad7e-266e6c10e531 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Deleting the datastore file [datastore1] a5c4afc6-38a6-4815-8ec4-cc01c24489bf {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 808.056203] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d27824cb-c678-446c-b852-c49020d83bab {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.065030] env[62385]: DEBUG oslo_vmware.api [None req-4e6ebd10-fb45-4b48-ad7e-266e6c10e531 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Waiting for the task: (returnval){ [ 808.065030] env[62385]: value = "task-1205779" [ 808.065030] env[62385]: _type = "Task" [ 808.065030] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.083394] env[62385]: DEBUG oslo_vmware.api [None req-4e6ebd10-fb45-4b48-ad7e-266e6c10e531 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205779, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.133616] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5d6a954b-460d-4b98-b7c6-e84c4438007a tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.313819] env[62385]: DEBUG oslo_vmware.api [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Task: {'id': task-1205774, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.415298] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205777, 'name': CreateVM_Task, 'duration_secs': 0.360978} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.415298] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 808.415854] env[62385]: DEBUG oslo_concurrency.lockutils [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.415919] env[62385]: DEBUG oslo_concurrency.lockutils [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.416304] env[62385]: DEBUG oslo_concurrency.lockutils [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 808.416550] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b234329-9324-4c37-a5a9-6b1827cbf4b5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.423581] env[62385]: DEBUG oslo_vmware.api [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Waiting for the task: (returnval){ [ 808.423581] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52de3fb9-d4cd-ce1c-9c4d-273cee9adbf6" [ 808.423581] env[62385]: _type = "Task" [ 808.423581] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.434168] env[62385]: DEBUG oslo_vmware.api [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52de3fb9-d4cd-ce1c-9c4d-273cee9adbf6, 'name': SearchDatastore_Task, 'duration_secs': 0.009101} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.434444] env[62385]: DEBUG oslo_concurrency.lockutils [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.434720] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 808.434953] env[62385]: DEBUG oslo_concurrency.lockutils [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.435073] env[62385]: DEBUG oslo_concurrency.lockutils [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.435263] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 808.435522] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dec212b7-f599-411a-af43-952992d03251 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.439294] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7633a67c-d409-4ad2-b6dd-1c2fb0a3aedf {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.443931] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 808.447176] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 808.447176] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb7f9b4f-f3b1-43a1-850e-14ce8d38035f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.448986] env[62385]: DEBUG oslo_concurrency.lockutils [None req-32376229-9c92-4dc6-bcbb-067cb39fb95b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "9d6f098a-0b05-43ef-96b0-9eb99ad3538c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 96.121s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.451007] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fb47a4f-cc78-41f3-b8d0-738f5c5796e0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.458409] env[62385]: DEBUG oslo_vmware.api [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Waiting for the task: (returnval){ [ 808.458409] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]525cd9ef-7679-7448-a94f-e894843ef9a7" [ 808.458409] env[62385]: _type = "Task" [ 808.458409] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.491468] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9527ffb9-0b13-49b3-93cf-9ac73264a5f2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.503677] env[62385]: DEBUG oslo_vmware.api [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]525cd9ef-7679-7448-a94f-e894843ef9a7, 'name': SearchDatastore_Task, 'duration_secs': 0.024181} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.505250] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-343ae3f5-1f13-4bf4-8a53-32a8b01a5c46 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.510283] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc67444c-84bb-49b5-b825-47c2ad905afa {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.517122] env[62385]: DEBUG oslo_vmware.api [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Waiting for the task: (returnval){ [ 808.517122] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]528ed2b1-57ac-4725-6a75-c3c2e853996d" [ 808.517122] env[62385]: _type = "Task" [ 808.517122] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.529464] env[62385]: DEBUG nova.compute.provider_tree [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 808.542483] env[62385]: DEBUG oslo_vmware.api [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]528ed2b1-57ac-4725-6a75-c3c2e853996d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.578971] env[62385]: DEBUG oslo_vmware.api [None req-4e6ebd10-fb45-4b48-ad7e-266e6c10e531 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Task: {'id': task-1205779, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.402512} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.579283] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e6ebd10-fb45-4b48-ad7e-266e6c10e531 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 808.579560] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4e6ebd10-fb45-4b48-ad7e-266e6c10e531 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 808.579754] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4e6ebd10-fb45-4b48-ad7e-266e6c10e531 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 808.579933] env[62385]: INFO nova.compute.manager [None req-4e6ebd10-fb45-4b48-ad7e-266e6c10e531 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Took 1.13 seconds to destroy the instance on the hypervisor. [ 808.580212] env[62385]: DEBUG oslo.service.loopingcall [None req-4e6ebd10-fb45-4b48-ad7e-266e6c10e531 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 808.580412] env[62385]: DEBUG nova.compute.manager [-] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 808.580511] env[62385]: DEBUG nova.network.neutron [-] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 808.682865] env[62385]: DEBUG nova.compute.manager [req-d0b593a5-6021-4c44-a750-937e6a2a4597 req-d5e34f87-6dd1-429e-9ba9-be563f5155fd service nova] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Received event network-vif-deleted-7965d15d-61f5-4fbb-8e38-53e0207a6c58 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 808.772913] env[62385]: DEBUG nova.network.neutron [req-d234aa25-e83a-4fbb-9eba-45c5ccd6117e req-c1729eb2-101d-4a79-a4b4-674944797b37 service nova] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Updated VIF entry in instance network info cache for port 32c6bd5f-7ef8-454d-8360-1bc0fcb5b894. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 808.773356] env[62385]: DEBUG nova.network.neutron [req-d234aa25-e83a-4fbb-9eba-45c5ccd6117e req-c1729eb2-101d-4a79-a4b4-674944797b37 service nova] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Updating instance_info_cache with network_info: [{"id": "32c6bd5f-7ef8-454d-8360-1bc0fcb5b894", "address": "fa:16:3e:82:b0:a8", "network": {"id": "5b5fe53c-e8fb-412e-b86d-a102146875c0", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.208", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "2e3b444561ba4b12b2f10b6744f6aa39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d5970ab5-34b8-4065-bfa6-f568b8f103b7", "external-id": "nsx-vlan-transportzone-418", "segmentation_id": 418, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32c6bd5f-7e", "ovs_interfaceid": "32c6bd5f-7ef8-454d-8360-1bc0fcb5b894", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.813935] env[62385]: DEBUG oslo_vmware.api [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Task: {'id': task-1205774, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.604478} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.814560] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 2b744879-01d2-440e-8adf-58438b922888/2b744879-01d2-440e-8adf-58438b922888.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 808.814794] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] [instance: 2b744879-01d2-440e-8adf-58438b922888] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 808.815122] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3763e156-6c62-4721-888a-d062204c242b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.822540] env[62385]: DEBUG oslo_vmware.api [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Waiting for the task: (returnval){ [ 808.822540] env[62385]: value = "task-1205780" [ 808.822540] env[62385]: _type = "Task" [ 808.822540] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.834241] env[62385]: DEBUG oslo_vmware.api [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Task: {'id': task-1205780, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.837819] env[62385]: DEBUG nova.compute.manager [req-9ea4e034-817e-4031-a7ca-742ad6ddf9d5 req-83da704d-d4b3-48c8-92c0-9fd3374be360 service nova] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Received event network-changed-ff03b686-6596-418e-b748-0a8a6799a2b5 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 808.838069] env[62385]: DEBUG nova.compute.manager [req-9ea4e034-817e-4031-a7ca-742ad6ddf9d5 req-83da704d-d4b3-48c8-92c0-9fd3374be360 service nova] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Refreshing instance network info cache due to event network-changed-ff03b686-6596-418e-b748-0a8a6799a2b5. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 808.838294] env[62385]: DEBUG oslo_concurrency.lockutils [req-9ea4e034-817e-4031-a7ca-742ad6ddf9d5 req-83da704d-d4b3-48c8-92c0-9fd3374be360 service nova] Acquiring lock "refresh_cache-9d6f098a-0b05-43ef-96b0-9eb99ad3538c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.838446] env[62385]: DEBUG oslo_concurrency.lockutils [req-9ea4e034-817e-4031-a7ca-742ad6ddf9d5 req-83da704d-d4b3-48c8-92c0-9fd3374be360 service nova] Acquired lock "refresh_cache-9d6f098a-0b05-43ef-96b0-9eb99ad3538c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.838635] env[62385]: DEBUG nova.network.neutron [req-9ea4e034-817e-4031-a7ca-742ad6ddf9d5 req-83da704d-d4b3-48c8-92c0-9fd3374be360 service nova] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Refreshing network info cache for port ff03b686-6596-418e-b748-0a8a6799a2b5 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 808.957787] env[62385]: DEBUG nova.compute.manager [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 809.037972] env[62385]: DEBUG nova.compute.manager [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 809.040517] env[62385]: DEBUG nova.scheduler.client.report [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 809.043771] env[62385]: DEBUG oslo_vmware.api [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]528ed2b1-57ac-4725-6a75-c3c2e853996d, 'name': SearchDatastore_Task, 'duration_secs': 0.025955} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.044656] env[62385]: DEBUG oslo_concurrency.lockutils [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.044754] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] e4089e71-eea1-4e16-b90c-97966333f2bc/e4089e71-eea1-4e16-b90c-97966333f2bc.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 809.044925] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-524a2836-3351-4653-bc6d-c7d60cda6746 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.053606] env[62385]: DEBUG oslo_vmware.api [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Waiting for the task: (returnval){ [ 809.053606] env[62385]: value = "task-1205781" [ 809.053606] env[62385]: _type = "Task" [ 809.053606] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.063768] env[62385]: DEBUG oslo_vmware.api [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205781, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.073618] env[62385]: DEBUG nova.virt.hardware [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 809.073960] env[62385]: DEBUG nova.virt.hardware [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 809.074234] env[62385]: DEBUG nova.virt.hardware [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 809.074485] env[62385]: DEBUG nova.virt.hardware [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 809.074729] env[62385]: DEBUG nova.virt.hardware [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 809.074956] env[62385]: DEBUG nova.virt.hardware [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 809.075329] env[62385]: DEBUG nova.virt.hardware [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 809.075582] env[62385]: DEBUG nova.virt.hardware [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 809.075840] env[62385]: DEBUG nova.virt.hardware [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 809.076081] env[62385]: DEBUG nova.virt.hardware [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 809.076401] env[62385]: DEBUG nova.virt.hardware [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 809.077389] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b8ef0cc-06b3-46cd-937d-48c548d82ca4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.088074] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-382b1301-5451-411f-8b0a-93db6b529eb7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.277035] env[62385]: DEBUG oslo_concurrency.lockutils [req-d234aa25-e83a-4fbb-9eba-45c5ccd6117e req-c1729eb2-101d-4a79-a4b4-674944797b37 service nova] Releasing lock "refresh_cache-e4089e71-eea1-4e16-b90c-97966333f2bc" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.332976] env[62385]: DEBUG oslo_vmware.api [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Task: {'id': task-1205780, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066708} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.333262] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] [instance: 2b744879-01d2-440e-8adf-58438b922888] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 809.335049] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edb4fad3-8684-4db0-96f6-81fd0d7715f0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.358163] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] [instance: 2b744879-01d2-440e-8adf-58438b922888] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] 2b744879-01d2-440e-8adf-58438b922888/2b744879-01d2-440e-8adf-58438b922888.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 809.359056] env[62385]: DEBUG nova.network.neutron [-] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.359454] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e428dfef-27f6-4bfe-80b1-8fd76c469e66 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.381756] env[62385]: DEBUG oslo_vmware.api [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Waiting for the task: (returnval){ [ 809.381756] env[62385]: value = "task-1205782" [ 809.381756] env[62385]: _type = "Task" [ 809.381756] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.390545] env[62385]: DEBUG oslo_vmware.api [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Task: {'id': task-1205782, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.479448] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.546180] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.548s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.547650] env[62385]: DEBUG nova.compute.manager [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 809.549387] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.635s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.553207] env[62385]: INFO nova.compute.claims [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 809.566158] env[62385]: DEBUG oslo_vmware.api [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205781, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.596131] env[62385]: DEBUG nova.network.neutron [req-9ea4e034-817e-4031-a7ca-742ad6ddf9d5 req-83da704d-d4b3-48c8-92c0-9fd3374be360 service nova] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Updated VIF entry in instance network info cache for port ff03b686-6596-418e-b748-0a8a6799a2b5. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 809.596577] env[62385]: DEBUG nova.network.neutron [req-9ea4e034-817e-4031-a7ca-742ad6ddf9d5 req-83da704d-d4b3-48c8-92c0-9fd3374be360 service nova] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Updating instance_info_cache with network_info: [{"id": "ff03b686-6596-418e-b748-0a8a6799a2b5", "address": "fa:16:3e:91:05:f2", "network": {"id": "afae6b07-d08c-413f-aa73-b545db1bb766", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1654965885-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.163", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "93111f5ba17c47de9691f90b99e03aa8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff03b686-65", "ovs_interfaceid": "ff03b686-6596-418e-b748-0a8a6799a2b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.611503] env[62385]: DEBUG nova.network.neutron [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Successfully updated port: 936e490f-ca20-418c-9b14-7e99db5ee2a9 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 809.874654] env[62385]: INFO nova.compute.manager [-] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Took 1.29 seconds to deallocate network for instance. [ 809.893975] env[62385]: DEBUG oslo_vmware.api [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Task: {'id': task-1205782, 'name': ReconfigVM_Task, 'duration_secs': 0.360926} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.894398] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] [instance: 2b744879-01d2-440e-8adf-58438b922888] Reconfigured VM instance instance-00000040 to attach disk [datastore2] 2b744879-01d2-440e-8adf-58438b922888/2b744879-01d2-440e-8adf-58438b922888.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 809.895229] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-84d28521-4668-4f02-a1e4-0f21ef559bc1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.903316] env[62385]: DEBUG oslo_vmware.api [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Waiting for the task: (returnval){ [ 809.903316] env[62385]: value = "task-1205783" [ 809.903316] env[62385]: _type = "Task" [ 809.903316] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.912766] env[62385]: DEBUG oslo_vmware.api [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Task: {'id': task-1205783, 'name': Rename_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.055698] env[62385]: DEBUG nova.compute.utils [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 810.069611] env[62385]: DEBUG nova.compute.manager [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 810.069611] env[62385]: DEBUG nova.network.neutron [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 810.084721] env[62385]: DEBUG oslo_vmware.api [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205781, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.90277} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.085018] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] e4089e71-eea1-4e16-b90c-97966333f2bc/e4089e71-eea1-4e16-b90c-97966333f2bc.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 810.085234] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 810.085495] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c516531d-632c-4ba3-81d5-c7bb2ca3fece {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.092777] env[62385]: DEBUG oslo_vmware.api [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Waiting for the task: (returnval){ [ 810.092777] env[62385]: value = "task-1205784" [ 810.092777] env[62385]: _type = "Task" [ 810.092777] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.099656] env[62385]: DEBUG oslo_concurrency.lockutils [req-9ea4e034-817e-4031-a7ca-742ad6ddf9d5 req-83da704d-d4b3-48c8-92c0-9fd3374be360 service nova] Releasing lock "refresh_cache-9d6f098a-0b05-43ef-96b0-9eb99ad3538c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.103234] env[62385]: DEBUG oslo_vmware.api [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205784, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.114026] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Acquiring lock "refresh_cache-f2be792c-a723-4a69-b255-bbe7f0e8692d" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.114217] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Acquired lock "refresh_cache-f2be792c-a723-4a69-b255-bbe7f0e8692d" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.114371] env[62385]: DEBUG nova.network.neutron [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 810.137160] env[62385]: DEBUG nova.policy [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '12f9521f9c9b4c6b9afc4973be32c5b7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '731b6700ce5b43369b40165545f04950', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 810.384892] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4e6ebd10-fb45-4b48-ad7e-266e6c10e531 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.415929] env[62385]: DEBUG oslo_vmware.api [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Task: {'id': task-1205783, 'name': Rename_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.486344] env[62385]: DEBUG nova.network.neutron [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Successfully created port: 5f63776d-8e86-4723-b873-9a8162076910 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 810.576380] env[62385]: DEBUG nova.compute.manager [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 810.604097] env[62385]: DEBUG oslo_vmware.api [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205784, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076269} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.604635] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 810.605274] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4be77014-e81b-46e5-a548-679770f35872 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.642409] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] e4089e71-eea1-4e16-b90c-97966333f2bc/e4089e71-eea1-4e16-b90c-97966333f2bc.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 810.642409] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e2b58569-7e65-4b1f-8b0f-60e9f0abade9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.666482] env[62385]: DEBUG oslo_vmware.api [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Waiting for the task: (returnval){ [ 810.666482] env[62385]: value = "task-1205785" [ 810.666482] env[62385]: _type = "Task" [ 810.666482] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.679695] env[62385]: DEBUG oslo_vmware.api [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205785, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.715014] env[62385]: DEBUG nova.compute.manager [req-0b359788-efdc-47bf-a689-28122ef8e805 req-05360aa6-769f-4800-96cd-5d1c4fa17307 service nova] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Received event network-vif-plugged-936e490f-ca20-418c-9b14-7e99db5ee2a9 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 810.715327] env[62385]: DEBUG oslo_concurrency.lockutils [req-0b359788-efdc-47bf-a689-28122ef8e805 req-05360aa6-769f-4800-96cd-5d1c4fa17307 service nova] Acquiring lock "f2be792c-a723-4a69-b255-bbe7f0e8692d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.715577] env[62385]: DEBUG oslo_concurrency.lockutils [req-0b359788-efdc-47bf-a689-28122ef8e805 req-05360aa6-769f-4800-96cd-5d1c4fa17307 service nova] Lock "f2be792c-a723-4a69-b255-bbe7f0e8692d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.715758] env[62385]: DEBUG oslo_concurrency.lockutils [req-0b359788-efdc-47bf-a689-28122ef8e805 req-05360aa6-769f-4800-96cd-5d1c4fa17307 service nova] Lock "f2be792c-a723-4a69-b255-bbe7f0e8692d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.715968] env[62385]: DEBUG nova.compute.manager [req-0b359788-efdc-47bf-a689-28122ef8e805 req-05360aa6-769f-4800-96cd-5d1c4fa17307 service nova] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] No waiting events found dispatching network-vif-plugged-936e490f-ca20-418c-9b14-7e99db5ee2a9 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 810.716195] env[62385]: WARNING nova.compute.manager [req-0b359788-efdc-47bf-a689-28122ef8e805 req-05360aa6-769f-4800-96cd-5d1c4fa17307 service nova] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Received unexpected event network-vif-plugged-936e490f-ca20-418c-9b14-7e99db5ee2a9 for instance with vm_state building and task_state spawning. [ 810.716487] env[62385]: DEBUG nova.compute.manager [req-0b359788-efdc-47bf-a689-28122ef8e805 req-05360aa6-769f-4800-96cd-5d1c4fa17307 service nova] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Received event network-changed-936e490f-ca20-418c-9b14-7e99db5ee2a9 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 810.716709] env[62385]: DEBUG nova.compute.manager [req-0b359788-efdc-47bf-a689-28122ef8e805 req-05360aa6-769f-4800-96cd-5d1c4fa17307 service nova] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Refreshing instance network info cache due to event network-changed-936e490f-ca20-418c-9b14-7e99db5ee2a9. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 810.716927] env[62385]: DEBUG oslo_concurrency.lockutils [req-0b359788-efdc-47bf-a689-28122ef8e805 req-05360aa6-769f-4800-96cd-5d1c4fa17307 service nova] Acquiring lock "refresh_cache-f2be792c-a723-4a69-b255-bbe7f0e8692d" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.739231] env[62385]: DEBUG nova.network.neutron [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 810.891452] env[62385]: DEBUG nova.compute.manager [req-ab240297-59c5-4da4-b54d-bdf714087e40 req-55c8bf62-7de2-4849-a21f-a2826131fa66 service nova] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Received event network-vif-deleted-81b5c1c2-0651-4034-87db-8faf51c3c7d5 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 810.892948] env[62385]: DEBUG nova.network.neutron [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Updating instance_info_cache with network_info: [{"id": "936e490f-ca20-418c-9b14-7e99db5ee2a9", "address": "fa:16:3e:8c:2d:e6", "network": {"id": "5b5fe53c-e8fb-412e-b86d-a102146875c0", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.193", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "2e3b444561ba4b12b2f10b6744f6aa39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d5970ab5-34b8-4065-bfa6-f568b8f103b7", "external-id": "nsx-vlan-transportzone-418", "segmentation_id": 418, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap936e490f-ca", "ovs_interfaceid": "936e490f-ca20-418c-9b14-7e99db5ee2a9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.918168] env[62385]: DEBUG oslo_vmware.api [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Task: {'id': task-1205783, 'name': Rename_Task, 'duration_secs': 0.916915} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.918509] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] [instance: 2b744879-01d2-440e-8adf-58438b922888] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 810.918775] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9a575cd4-cdb0-4f80-8749-2e35fda11838 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.926404] env[62385]: DEBUG oslo_vmware.api [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Waiting for the task: (returnval){ [ 810.926404] env[62385]: value = "task-1205786" [ 810.926404] env[62385]: _type = "Task" [ 810.926404] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.937319] env[62385]: DEBUG oslo_vmware.api [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Task: {'id': task-1205786, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.002161] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0784594e-4bef-4c09-bbcf-90ab947e4869 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.009735] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d5f13d8-e845-44af-a96f-b2c26989d04d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.043837] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c038b91-5b45-4fec-ae19-eedf9b6f1980 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.053031] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72ffe560-07e9-486b-bbeb-07cdecf841e5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.070038] env[62385]: DEBUG nova.compute.provider_tree [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 811.178012] env[62385]: DEBUG oslo_vmware.api [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205785, 'name': ReconfigVM_Task, 'duration_secs': 0.400382} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.178336] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Reconfigured VM instance instance-00000041 to attach disk [datastore2] e4089e71-eea1-4e16-b90c-97966333f2bc/e4089e71-eea1-4e16-b90c-97966333f2bc.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 811.179040] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c9f5c208-6e81-4297-af6c-f28285146753 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.189076] env[62385]: DEBUG oslo_vmware.api [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Waiting for the task: (returnval){ [ 811.189076] env[62385]: value = "task-1205787" [ 811.189076] env[62385]: _type = "Task" [ 811.189076] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.201460] env[62385]: DEBUG oslo_vmware.api [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205787, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.397889] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Releasing lock "refresh_cache-f2be792c-a723-4a69-b255-bbe7f0e8692d" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.398203] env[62385]: DEBUG nova.compute.manager [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Instance network_info: |[{"id": "936e490f-ca20-418c-9b14-7e99db5ee2a9", "address": "fa:16:3e:8c:2d:e6", "network": {"id": "5b5fe53c-e8fb-412e-b86d-a102146875c0", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.193", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "2e3b444561ba4b12b2f10b6744f6aa39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d5970ab5-34b8-4065-bfa6-f568b8f103b7", "external-id": "nsx-vlan-transportzone-418", "segmentation_id": 418, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap936e490f-ca", "ovs_interfaceid": "936e490f-ca20-418c-9b14-7e99db5ee2a9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 811.399117] env[62385]: DEBUG oslo_concurrency.lockutils [req-0b359788-efdc-47bf-a689-28122ef8e805 req-05360aa6-769f-4800-96cd-5d1c4fa17307 service nova] Acquired lock "refresh_cache-f2be792c-a723-4a69-b255-bbe7f0e8692d" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.399117] env[62385]: DEBUG nova.network.neutron [req-0b359788-efdc-47bf-a689-28122ef8e805 req-05360aa6-769f-4800-96cd-5d1c4fa17307 service nova] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Refreshing network info cache for port 936e490f-ca20-418c-9b14-7e99db5ee2a9 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 811.400428] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8c:2d:e6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd5970ab5-34b8-4065-bfa6-f568b8f103b7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '936e490f-ca20-418c-9b14-7e99db5ee2a9', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 811.408480] env[62385]: DEBUG oslo.service.loopingcall [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 811.409083] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 811.409334] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8d777bb7-fe68-4737-8811-1c9e06393a93 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.432711] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 811.432711] env[62385]: value = "task-1205788" [ 811.432711] env[62385]: _type = "Task" [ 811.432711] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.439371] env[62385]: DEBUG oslo_vmware.api [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Task: {'id': task-1205786, 'name': PowerOnVM_Task, 'duration_secs': 0.501984} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.439997] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] [instance: 2b744879-01d2-440e-8adf-58438b922888] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 811.440244] env[62385]: INFO nova.compute.manager [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] [instance: 2b744879-01d2-440e-8adf-58438b922888] Took 9.51 seconds to spawn the instance on the hypervisor. [ 811.440477] env[62385]: DEBUG nova.compute.manager [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] [instance: 2b744879-01d2-440e-8adf-58438b922888] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 811.441733] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f3141f2-154b-4dbe-80c0-9c6fd215a4ed {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.447347] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205788, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.577577] env[62385]: DEBUG nova.scheduler.client.report [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 811.585703] env[62385]: DEBUG nova.compute.manager [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 811.611648] env[62385]: DEBUG nova.virt.hardware [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 811.611902] env[62385]: DEBUG nova.virt.hardware [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 811.612080] env[62385]: DEBUG nova.virt.hardware [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 811.612274] env[62385]: DEBUG nova.virt.hardware [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 811.612489] env[62385]: DEBUG nova.virt.hardware [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 811.612669] env[62385]: DEBUG nova.virt.hardware [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 811.612889] env[62385]: DEBUG nova.virt.hardware [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 811.613073] env[62385]: DEBUG nova.virt.hardware [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 811.613256] env[62385]: DEBUG nova.virt.hardware [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 811.613427] env[62385]: DEBUG nova.virt.hardware [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 811.613605] env[62385]: DEBUG nova.virt.hardware [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 811.614689] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1f846ab-f3ba-4e23-97ba-08a85ca0ffd8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.624660] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27255d21-03c2-4c3e-baff-91fc979b277d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.698402] env[62385]: DEBUG oslo_vmware.api [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205787, 'name': Rename_Task, 'duration_secs': 0.178138} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.698716] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 811.698970] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-03e24abe-14e7-43cc-aab6-204d02abfe2a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.712580] env[62385]: DEBUG oslo_vmware.api [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Waiting for the task: (returnval){ [ 811.712580] env[62385]: value = "task-1205789" [ 811.712580] env[62385]: _type = "Task" [ 811.712580] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.721193] env[62385]: DEBUG oslo_vmware.api [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205789, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.948370] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205788, 'name': CreateVM_Task, 'duration_secs': 0.443213} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.948370] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 811.948370] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.948370] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.948370] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 811.948370] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6f85cc4-db0e-45d7-8a77-f7e3bc0b2c2b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.953078] env[62385]: DEBUG oslo_vmware.api [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Waiting for the task: (returnval){ [ 811.953078] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52b32edf-edce-dc62-cc3d-9e7aee2ba32b" [ 811.953078] env[62385]: _type = "Task" [ 811.953078] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.968928] env[62385]: DEBUG oslo_vmware.api [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52b32edf-edce-dc62-cc3d-9e7aee2ba32b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.971195] env[62385]: INFO nova.compute.manager [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] [instance: 2b744879-01d2-440e-8adf-58438b922888] Took 41.61 seconds to build instance. [ 811.976180] env[62385]: DEBUG nova.compute.manager [req-7da2d714-a57a-4951-93f1-019cd5633057 req-e515de2c-df4a-46f9-a8b6-4a1e6919919e service nova] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Received event network-vif-plugged-5f63776d-8e86-4723-b873-9a8162076910 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 811.976180] env[62385]: DEBUG oslo_concurrency.lockutils [req-7da2d714-a57a-4951-93f1-019cd5633057 req-e515de2c-df4a-46f9-a8b6-4a1e6919919e service nova] Acquiring lock "55d62bba-ff30-42bf-b1cb-7567988a9361-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.976392] env[62385]: DEBUG oslo_concurrency.lockutils [req-7da2d714-a57a-4951-93f1-019cd5633057 req-e515de2c-df4a-46f9-a8b6-4a1e6919919e service nova] Lock "55d62bba-ff30-42bf-b1cb-7567988a9361-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.976574] env[62385]: DEBUG oslo_concurrency.lockutils [req-7da2d714-a57a-4951-93f1-019cd5633057 req-e515de2c-df4a-46f9-a8b6-4a1e6919919e service nova] Lock "55d62bba-ff30-42bf-b1cb-7567988a9361-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.976744] env[62385]: DEBUG nova.compute.manager [req-7da2d714-a57a-4951-93f1-019cd5633057 req-e515de2c-df4a-46f9-a8b6-4a1e6919919e service nova] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] No waiting events found dispatching network-vif-plugged-5f63776d-8e86-4723-b873-9a8162076910 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 811.976927] env[62385]: WARNING nova.compute.manager [req-7da2d714-a57a-4951-93f1-019cd5633057 req-e515de2c-df4a-46f9-a8b6-4a1e6919919e service nova] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Received unexpected event network-vif-plugged-5f63776d-8e86-4723-b873-9a8162076910 for instance with vm_state building and task_state spawning. [ 812.082945] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.534s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.083541] env[62385]: DEBUG nova.compute.manager [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 812.086432] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b6753bc4-02bf-4300-903f-acd56a4146b9 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.217s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.086543] env[62385]: DEBUG nova.objects.instance [None req-b6753bc4-02bf-4300-903f-acd56a4146b9 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Lazy-loading 'resources' on Instance uuid e73ecefc-fed3-4ccd-88d3-399a4b72bb4b {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 812.162676] env[62385]: DEBUG nova.network.neutron [req-0b359788-efdc-47bf-a689-28122ef8e805 req-05360aa6-769f-4800-96cd-5d1c4fa17307 service nova] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Updated VIF entry in instance network info cache for port 936e490f-ca20-418c-9b14-7e99db5ee2a9. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 812.163054] env[62385]: DEBUG nova.network.neutron [req-0b359788-efdc-47bf-a689-28122ef8e805 req-05360aa6-769f-4800-96cd-5d1c4fa17307 service nova] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Updating instance_info_cache with network_info: [{"id": "936e490f-ca20-418c-9b14-7e99db5ee2a9", "address": "fa:16:3e:8c:2d:e6", "network": {"id": "5b5fe53c-e8fb-412e-b86d-a102146875c0", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.193", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "2e3b444561ba4b12b2f10b6744f6aa39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d5970ab5-34b8-4065-bfa6-f568b8f103b7", "external-id": "nsx-vlan-transportzone-418", "segmentation_id": 418, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap936e490f-ca", "ovs_interfaceid": "936e490f-ca20-418c-9b14-7e99db5ee2a9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.224697] env[62385]: DEBUG oslo_vmware.api [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205789, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.464938] env[62385]: DEBUG oslo_vmware.api [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52b32edf-edce-dc62-cc3d-9e7aee2ba32b, 'name': SearchDatastore_Task, 'duration_secs': 0.012933} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.465308] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.465590] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 812.465895] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.466117] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.466331] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 812.466656] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b02c2b2f-7422-4a0a-b891-87bb47583cd2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.473359] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9655ac5f-2054-493f-8c93-a63da4ad59f2 tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Lock "2b744879-01d2-440e-8adf-58438b922888" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 99.400s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.476780] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 812.476964] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 812.477757] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b3fbbc19-2181-410d-a512-a842a84466f2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.484353] env[62385]: DEBUG oslo_vmware.api [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Waiting for the task: (returnval){ [ 812.484353] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]526ee644-36b7-b45f-8997-8c6eae62bced" [ 812.484353] env[62385]: _type = "Task" [ 812.484353] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.494008] env[62385]: DEBUG oslo_vmware.api [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]526ee644-36b7-b45f-8997-8c6eae62bced, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.528093] env[62385]: DEBUG nova.network.neutron [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Successfully updated port: 5f63776d-8e86-4723-b873-9a8162076910 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 812.595049] env[62385]: DEBUG nova.compute.utils [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 812.597651] env[62385]: DEBUG nova.compute.manager [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 812.597768] env[62385]: DEBUG nova.network.neutron [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 812.640524] env[62385]: DEBUG nova.policy [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7e04c6591fe043d686140eb2f5ce0be2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd45b5346eab44707b4d81eaaf12b0935', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 812.665850] env[62385]: DEBUG oslo_concurrency.lockutils [req-0b359788-efdc-47bf-a689-28122ef8e805 req-05360aa6-769f-4800-96cd-5d1c4fa17307 service nova] Releasing lock "refresh_cache-f2be792c-a723-4a69-b255-bbe7f0e8692d" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.725824] env[62385]: DEBUG oslo_vmware.api [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205789, 'name': PowerOnVM_Task, 'duration_secs': 0.915786} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.728478] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 812.728663] env[62385]: INFO nova.compute.manager [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Took 8.28 seconds to spawn the instance on the hypervisor. [ 812.728885] env[62385]: DEBUG nova.compute.manager [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 812.729844] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8260b0f-9925-4dd8-9be8-6189c451ec69 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.746019] env[62385]: DEBUG nova.compute.manager [req-f0db8964-0817-42cf-9f14-0b85109913f7 req-10fddf73-e7bf-4500-a32e-45436668110b service nova] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Received event network-changed-5f63776d-8e86-4723-b873-9a8162076910 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 812.746019] env[62385]: DEBUG nova.compute.manager [req-f0db8964-0817-42cf-9f14-0b85109913f7 req-10fddf73-e7bf-4500-a32e-45436668110b service nova] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Refreshing instance network info cache due to event network-changed-5f63776d-8e86-4723-b873-9a8162076910. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 812.746019] env[62385]: DEBUG oslo_concurrency.lockutils [req-f0db8964-0817-42cf-9f14-0b85109913f7 req-10fddf73-e7bf-4500-a32e-45436668110b service nova] Acquiring lock "refresh_cache-55d62bba-ff30-42bf-b1cb-7567988a9361" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.746019] env[62385]: DEBUG oslo_concurrency.lockutils [req-f0db8964-0817-42cf-9f14-0b85109913f7 req-10fddf73-e7bf-4500-a32e-45436668110b service nova] Acquired lock "refresh_cache-55d62bba-ff30-42bf-b1cb-7567988a9361" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.746019] env[62385]: DEBUG nova.network.neutron [req-f0db8964-0817-42cf-9f14-0b85109913f7 req-10fddf73-e7bf-4500-a32e-45436668110b service nova] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Refreshing network info cache for port 5f63776d-8e86-4723-b873-9a8162076910 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 812.927140] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-373b7fa6-ccb1-47dc-aa52-c362430ca425 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.935484] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f178ef76-86b7-4883-9ac4-a156553a0e79 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.964959] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0360d6eb-89dd-45d0-aa7b-42b20d756343 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.974412] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-877e334f-4b05-42f9-a1cd-26877a741623 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.978776] env[62385]: DEBUG nova.compute.manager [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 812.982332] env[62385]: DEBUG nova.network.neutron [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Successfully created port: 404c104c-258f-480e-bfe2-4ace4e83a43c {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 812.998684] env[62385]: DEBUG nova.compute.provider_tree [None req-b6753bc4-02bf-4300-903f-acd56a4146b9 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 813.005915] env[62385]: DEBUG oslo_vmware.api [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]526ee644-36b7-b45f-8997-8c6eae62bced, 'name': SearchDatastore_Task, 'duration_secs': 0.014718} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.006707] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1f40c7b-130d-43cb-987c-fc474d315ef2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.013877] env[62385]: DEBUG oslo_vmware.api [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Waiting for the task: (returnval){ [ 813.013877] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52d5ce47-4d76-6761-4bad-c183ab15f31b" [ 813.013877] env[62385]: _type = "Task" [ 813.013877] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.022425] env[62385]: DEBUG oslo_vmware.api [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52d5ce47-4d76-6761-4bad-c183ab15f31b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.030356] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquiring lock "refresh_cache-55d62bba-ff30-42bf-b1cb-7567988a9361" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 813.100621] env[62385]: DEBUG nova.compute.manager [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 813.255985] env[62385]: INFO nova.compute.manager [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Took 39.00 seconds to build instance. [ 813.287726] env[62385]: DEBUG nova.network.neutron [req-f0db8964-0817-42cf-9f14-0b85109913f7 req-10fddf73-e7bf-4500-a32e-45436668110b service nova] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 813.357203] env[62385]: DEBUG nova.network.neutron [req-f0db8964-0817-42cf-9f14-0b85109913f7 req-10fddf73-e7bf-4500-a32e-45436668110b service nova] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.502624] env[62385]: DEBUG nova.scheduler.client.report [None req-b6753bc4-02bf-4300-903f-acd56a4146b9 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 813.512669] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.525831] env[62385]: DEBUG oslo_vmware.api [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52d5ce47-4d76-6761-4bad-c183ab15f31b, 'name': SearchDatastore_Task, 'duration_secs': 0.009706} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.527227] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.527597] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] f2be792c-a723-4a69-b255-bbe7f0e8692d/f2be792c-a723-4a69-b255-bbe7f0e8692d.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 813.528300] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-91c4b77a-bf9b-4950-a9ce-6edc7b192597 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.538357] env[62385]: DEBUG oslo_vmware.api [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Waiting for the task: (returnval){ [ 813.538357] env[62385]: value = "task-1205790" [ 813.538357] env[62385]: _type = "Task" [ 813.538357] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.548784] env[62385]: DEBUG oslo_vmware.api [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205790, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.719345] env[62385]: DEBUG oslo_concurrency.lockutils [None req-695d2e36-0194-441c-bb9f-b1d1d915536f tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Acquiring lock "2b744879-01d2-440e-8adf-58438b922888" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.719345] env[62385]: DEBUG oslo_concurrency.lockutils [None req-695d2e36-0194-441c-bb9f-b1d1d915536f tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Lock "2b744879-01d2-440e-8adf-58438b922888" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.719345] env[62385]: DEBUG oslo_concurrency.lockutils [None req-695d2e36-0194-441c-bb9f-b1d1d915536f tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Acquiring lock "2b744879-01d2-440e-8adf-58438b922888-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.719345] env[62385]: DEBUG oslo_concurrency.lockutils [None req-695d2e36-0194-441c-bb9f-b1d1d915536f tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Lock "2b744879-01d2-440e-8adf-58438b922888-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.719745] env[62385]: DEBUG oslo_concurrency.lockutils [None req-695d2e36-0194-441c-bb9f-b1d1d915536f tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Lock "2b744879-01d2-440e-8adf-58438b922888-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.721523] env[62385]: INFO nova.compute.manager [None req-695d2e36-0194-441c-bb9f-b1d1d915536f tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] [instance: 2b744879-01d2-440e-8adf-58438b922888] Terminating instance [ 813.723608] env[62385]: DEBUG nova.compute.manager [None req-695d2e36-0194-441c-bb9f-b1d1d915536f tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] [instance: 2b744879-01d2-440e-8adf-58438b922888] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 813.723833] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-695d2e36-0194-441c-bb9f-b1d1d915536f tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] [instance: 2b744879-01d2-440e-8adf-58438b922888] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 813.725480] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dc2b8b4-fcb5-4ced-b61a-ff7637926a8e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.735380] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-695d2e36-0194-441c-bb9f-b1d1d915536f tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] [instance: 2b744879-01d2-440e-8adf-58438b922888] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 813.735747] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a9d05133-5e67-439f-a172-404ae5e6296b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.746638] env[62385]: DEBUG oslo_vmware.api [None req-695d2e36-0194-441c-bb9f-b1d1d915536f tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Waiting for the task: (returnval){ [ 813.746638] env[62385]: value = "task-1205791" [ 813.746638] env[62385]: _type = "Task" [ 813.746638] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.760603] env[62385]: DEBUG oslo_concurrency.lockutils [None req-dd2e4b75-99fd-4c5b-94e1-2ac1f54afb23 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Lock "e4089e71-eea1-4e16-b90c-97966333f2bc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 86.762s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.760960] env[62385]: DEBUG oslo_vmware.api [None req-695d2e36-0194-441c-bb9f-b1d1d915536f tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Task: {'id': task-1205791, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.860879] env[62385]: DEBUG oslo_concurrency.lockutils [req-f0db8964-0817-42cf-9f14-0b85109913f7 req-10fddf73-e7bf-4500-a32e-45436668110b service nova] Releasing lock "refresh_cache-55d62bba-ff30-42bf-b1cb-7567988a9361" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.861022] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquired lock "refresh_cache-55d62bba-ff30-42bf-b1cb-7567988a9361" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.861139] env[62385]: DEBUG nova.network.neutron [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 814.013016] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b6753bc4-02bf-4300-903f-acd56a4146b9 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.926s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.016368] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d08062cc-cece-446b-835d-8aef2e093696 tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.455s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.019119] env[62385]: DEBUG nova.objects.instance [None req-d08062cc-cece-446b-835d-8aef2e093696 tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Lazy-loading 'resources' on Instance uuid f77dc5ab-7e63-4a1b-9c87-e474a7e28c93 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 814.036574] env[62385]: INFO nova.scheduler.client.report [None req-b6753bc4-02bf-4300-903f-acd56a4146b9 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Deleted allocations for instance e73ecefc-fed3-4ccd-88d3-399a4b72bb4b [ 814.051127] env[62385]: DEBUG oslo_vmware.api [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205790, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.111293] env[62385]: DEBUG nova.compute.manager [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 814.136955] env[62385]: DEBUG nova.virt.hardware [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 814.137217] env[62385]: DEBUG nova.virt.hardware [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 814.137381] env[62385]: DEBUG nova.virt.hardware [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 814.137570] env[62385]: DEBUG nova.virt.hardware [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 814.137764] env[62385]: DEBUG nova.virt.hardware [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 814.137868] env[62385]: DEBUG nova.virt.hardware [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 814.138097] env[62385]: DEBUG nova.virt.hardware [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 814.138268] env[62385]: DEBUG nova.virt.hardware [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 814.138472] env[62385]: DEBUG nova.virt.hardware [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 814.138655] env[62385]: DEBUG nova.virt.hardware [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 814.138833] env[62385]: DEBUG nova.virt.hardware [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 814.139728] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4bab228-c0d9-442f-a6e3-9ed354f00ccc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.147935] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26329768-ef3b-406f-b92d-ccb30a01ed12 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.258202] env[62385]: DEBUG oslo_vmware.api [None req-695d2e36-0194-441c-bb9f-b1d1d915536f tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Task: {'id': task-1205791, 'name': PowerOffVM_Task, 'duration_secs': 0.278354} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.258526] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-695d2e36-0194-441c-bb9f-b1d1d915536f tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] [instance: 2b744879-01d2-440e-8adf-58438b922888] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 814.258712] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-695d2e36-0194-441c-bb9f-b1d1d915536f tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] [instance: 2b744879-01d2-440e-8adf-58438b922888] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 814.258961] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5820dad9-b32d-4aab-a648-2886f47fe784 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.321842] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-695d2e36-0194-441c-bb9f-b1d1d915536f tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] [instance: 2b744879-01d2-440e-8adf-58438b922888] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 814.322092] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-695d2e36-0194-441c-bb9f-b1d1d915536f tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] [instance: 2b744879-01d2-440e-8adf-58438b922888] Deleting contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 814.322289] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-695d2e36-0194-441c-bb9f-b1d1d915536f tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Deleting the datastore file [datastore2] 2b744879-01d2-440e-8adf-58438b922888 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 814.322558] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c1879dce-6c47-420d-b1f5-59e31e9062d1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.328680] env[62385]: DEBUG oslo_vmware.api [None req-695d2e36-0194-441c-bb9f-b1d1d915536f tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Waiting for the task: (returnval){ [ 814.328680] env[62385]: value = "task-1205793" [ 814.328680] env[62385]: _type = "Task" [ 814.328680] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.336468] env[62385]: DEBUG oslo_vmware.api [None req-695d2e36-0194-441c-bb9f-b1d1d915536f tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Task: {'id': task-1205793, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.394669] env[62385]: DEBUG nova.network.neutron [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 814.535691] env[62385]: DEBUG nova.network.neutron [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Updating instance_info_cache with network_info: [{"id": "5f63776d-8e86-4723-b873-9a8162076910", "address": "fa:16:3e:2f:b9:ef", "network": {"id": "2c1e2031-ab59-47f5-a758-3a314aca6c1e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1100903482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "731b6700ce5b43369b40165545f04950", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1316f5aa-529f-4bac-8dd7-6076a9d43312", "external-id": "nsx-vlan-transportzone-399", "segmentation_id": 399, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f63776d-8e", "ovs_interfaceid": "5f63776d-8e86-4723-b873-9a8162076910", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 814.551703] env[62385]: DEBUG oslo_vmware.api [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205790, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.555018] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b6753bc4-02bf-4300-903f-acd56a4146b9 tempest-SecurityGroupsTestJSON-827094486 tempest-SecurityGroupsTestJSON-827094486-project-member] Lock "e73ecefc-fed3-4ccd-88d3-399a4b72bb4b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.360s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.785986] env[62385]: DEBUG nova.network.neutron [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Successfully updated port: 404c104c-258f-480e-bfe2-4ace4e83a43c {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 814.842148] env[62385]: DEBUG oslo_vmware.api [None req-695d2e36-0194-441c-bb9f-b1d1d915536f tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Task: {'id': task-1205793, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.846524] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b5659be-a17d-41a1-afb7-35102adb5225 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.853777] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df9d38e1-4825-4e32-8592-e33c635e5829 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.886447] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f595736e-cd29-48f7-a3cd-2fe166d88dd7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.894759] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30e1989d-5edd-4594-b3e0-ec1e98a54141 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.912640] env[62385]: DEBUG nova.compute.provider_tree [None req-d08062cc-cece-446b-835d-8aef2e093696 tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 815.001920] env[62385]: DEBUG nova.compute.manager [req-8ba7b730-e26b-478c-a07b-3377b1a08799 req-273e3dc2-1a7c-44bf-a620-e1b6b451c59c service nova] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Received event network-vif-plugged-404c104c-258f-480e-bfe2-4ace4e83a43c {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 815.002208] env[62385]: DEBUG oslo_concurrency.lockutils [req-8ba7b730-e26b-478c-a07b-3377b1a08799 req-273e3dc2-1a7c-44bf-a620-e1b6b451c59c service nova] Acquiring lock "9b9fd167-d843-4b9d-9e4f-dbde5d3d629e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.002462] env[62385]: DEBUG oslo_concurrency.lockutils [req-8ba7b730-e26b-478c-a07b-3377b1a08799 req-273e3dc2-1a7c-44bf-a620-e1b6b451c59c service nova] Lock "9b9fd167-d843-4b9d-9e4f-dbde5d3d629e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.002647] env[62385]: DEBUG oslo_concurrency.lockutils [req-8ba7b730-e26b-478c-a07b-3377b1a08799 req-273e3dc2-1a7c-44bf-a620-e1b6b451c59c service nova] Lock "9b9fd167-d843-4b9d-9e4f-dbde5d3d629e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.002820] env[62385]: DEBUG nova.compute.manager [req-8ba7b730-e26b-478c-a07b-3377b1a08799 req-273e3dc2-1a7c-44bf-a620-e1b6b451c59c service nova] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] No waiting events found dispatching network-vif-plugged-404c104c-258f-480e-bfe2-4ace4e83a43c {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 815.002991] env[62385]: WARNING nova.compute.manager [req-8ba7b730-e26b-478c-a07b-3377b1a08799 req-273e3dc2-1a7c-44bf-a620-e1b6b451c59c service nova] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Received unexpected event network-vif-plugged-404c104c-258f-480e-bfe2-4ace4e83a43c for instance with vm_state building and task_state spawning. [ 815.036875] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Releasing lock "refresh_cache-55d62bba-ff30-42bf-b1cb-7567988a9361" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.037205] env[62385]: DEBUG nova.compute.manager [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Instance network_info: |[{"id": "5f63776d-8e86-4723-b873-9a8162076910", "address": "fa:16:3e:2f:b9:ef", "network": {"id": "2c1e2031-ab59-47f5-a758-3a314aca6c1e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1100903482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "731b6700ce5b43369b40165545f04950", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1316f5aa-529f-4bac-8dd7-6076a9d43312", "external-id": "nsx-vlan-transportzone-399", "segmentation_id": 399, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f63776d-8e", "ovs_interfaceid": "5f63776d-8e86-4723-b873-9a8162076910", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 815.037700] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2f:b9:ef', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1316f5aa-529f-4bac-8dd7-6076a9d43312', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5f63776d-8e86-4723-b873-9a8162076910', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 815.045586] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Creating folder: Project (731b6700ce5b43369b40165545f04950). Parent ref: group-v261107. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 815.046104] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e0e23d43-d52a-45f1-a8a0-fe8d5c0ff328 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.058719] env[62385]: DEBUG oslo_vmware.api [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205790, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.069964] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Created folder: Project (731b6700ce5b43369b40165545f04950) in parent group-v261107. [ 815.070249] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Creating folder: Instances. Parent ref: group-v261199. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 815.071143] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-619e9abf-6cf6-42ee-950c-c94b880f50ae {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.085705] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Created folder: Instances in parent group-v261199. [ 815.089023] env[62385]: DEBUG oslo.service.loopingcall [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 815.089023] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 815.089023] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f923a488-f9e8-42db-818b-ae985f00049e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.113091] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 815.113091] env[62385]: value = "task-1205796" [ 815.113091] env[62385]: _type = "Task" [ 815.113091] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.131711] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205796, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.291191] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "refresh_cache-9b9fd167-d843-4b9d-9e4f-dbde5d3d629e" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.291371] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquired lock "refresh_cache-9b9fd167-d843-4b9d-9e4f-dbde5d3d629e" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.291371] env[62385]: DEBUG nova.network.neutron [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 815.341472] env[62385]: DEBUG oslo_vmware.api [None req-695d2e36-0194-441c-bb9f-b1d1d915536f tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Task: {'id': task-1205793, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.770321} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.341829] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-695d2e36-0194-441c-bb9f-b1d1d915536f tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 815.342040] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-695d2e36-0194-441c-bb9f-b1d1d915536f tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] [instance: 2b744879-01d2-440e-8adf-58438b922888] Deleted contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 815.342989] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-695d2e36-0194-441c-bb9f-b1d1d915536f tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] [instance: 2b744879-01d2-440e-8adf-58438b922888] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 815.343219] env[62385]: INFO nova.compute.manager [None req-695d2e36-0194-441c-bb9f-b1d1d915536f tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] [instance: 2b744879-01d2-440e-8adf-58438b922888] Took 1.62 seconds to destroy the instance on the hypervisor. [ 815.343485] env[62385]: DEBUG oslo.service.loopingcall [None req-695d2e36-0194-441c-bb9f-b1d1d915536f tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 815.344057] env[62385]: DEBUG nova.compute.manager [-] [instance: 2b744879-01d2-440e-8adf-58438b922888] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 815.344167] env[62385]: DEBUG nova.network.neutron [-] [instance: 2b744879-01d2-440e-8adf-58438b922888] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 815.418939] env[62385]: DEBUG nova.scheduler.client.report [None req-d08062cc-cece-446b-835d-8aef2e093696 tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 815.558558] env[62385]: DEBUG oslo_vmware.api [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205790, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.525775} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.559018] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] f2be792c-a723-4a69-b255-bbe7f0e8692d/f2be792c-a723-4a69-b255-bbe7f0e8692d.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 815.559352] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 815.559729] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5532999f-1259-4f10-9d4b-086227b9a825 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.568616] env[62385]: DEBUG oslo_vmware.api [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Waiting for the task: (returnval){ [ 815.568616] env[62385]: value = "task-1205797" [ 815.568616] env[62385]: _type = "Task" [ 815.568616] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.578046] env[62385]: DEBUG oslo_vmware.api [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205797, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.626272] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205796, 'name': CreateVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.822832] env[62385]: DEBUG nova.network.neutron [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 815.925706] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d08062cc-cece-446b-835d-8aef2e093696 tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.909s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.932783] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.763s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.937412] env[62385]: INFO nova.compute.claims [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 815.961744] env[62385]: INFO nova.scheduler.client.report [None req-d08062cc-cece-446b-835d-8aef2e093696 tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Deleted allocations for instance f77dc5ab-7e63-4a1b-9c87-e474a7e28c93 [ 815.996744] env[62385]: DEBUG nova.network.neutron [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Updating instance_info_cache with network_info: [{"id": "404c104c-258f-480e-bfe2-4ace4e83a43c", "address": "fa:16:3e:87:f8:2a", "network": {"id": "06017161-984c-4657-9639-4e6bdb6d4c09", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1072882119-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d45b5346eab44707b4d81eaaf12b0935", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39cd75b0-9ec7-48ed-b57f-34da0c573a60", "external-id": "nsx-vlan-transportzone-751", "segmentation_id": 751, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap404c104c-25", "ovs_interfaceid": "404c104c-258f-480e-bfe2-4ace4e83a43c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.079766] env[62385]: DEBUG oslo_vmware.api [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205797, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064288} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.080099] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 816.080920] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-846d5484-c17f-4257-bc30-3f21fd0585d9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.105645] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] f2be792c-a723-4a69-b255-bbe7f0e8692d/f2be792c-a723-4a69-b255-bbe7f0e8692d.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 816.106654] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ec885972-8150-41a4-88ec-b59d836c2600 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.129837] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205796, 'name': CreateVM_Task, 'duration_secs': 0.60862} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.130986] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 816.131360] env[62385]: DEBUG oslo_vmware.api [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Waiting for the task: (returnval){ [ 816.131360] env[62385]: value = "task-1205798" [ 816.131360] env[62385]: _type = "Task" [ 816.131360] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.132015] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.132201] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.132535] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 816.132831] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f262b632-ce3b-4444-a1a9-1688a884cfc8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.140807] env[62385]: DEBUG oslo_vmware.api [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 816.140807] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c601df-d1f3-f3c1-8bbb-d1cbd89a8e89" [ 816.140807] env[62385]: _type = "Task" [ 816.140807] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.144715] env[62385]: DEBUG oslo_vmware.api [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205798, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.155174] env[62385]: DEBUG oslo_vmware.api [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c601df-d1f3-f3c1-8bbb-d1cbd89a8e89, 'name': SearchDatastore_Task, 'duration_secs': 0.010679} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.155174] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.155422] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 816.155694] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.155847] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.156051] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 816.156351] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8a3b1045-94b2-43fe-bc10-e757b2392aa9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.165793] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 816.166007] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 816.166822] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27943a1b-7f29-4fde-b827-7df1fd28cdd5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.173166] env[62385]: DEBUG oslo_vmware.api [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 816.173166] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52114399-8e43-161d-2d5a-b4f0990c647b" [ 816.173166] env[62385]: _type = "Task" [ 816.173166] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.181984] env[62385]: DEBUG oslo_vmware.api [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52114399-8e43-161d-2d5a-b4f0990c647b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.470277] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d08062cc-cece-446b-835d-8aef2e093696 tempest-ServerAddressesTestJSON-2133113904 tempest-ServerAddressesTestJSON-2133113904-project-member] Lock "f77dc5ab-7e63-4a1b-9c87-e474a7e28c93" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.231s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.480163] env[62385]: DEBUG nova.network.neutron [-] [instance: 2b744879-01d2-440e-8adf-58438b922888] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.501291] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Releasing lock "refresh_cache-9b9fd167-d843-4b9d-9e4f-dbde5d3d629e" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.501291] env[62385]: DEBUG nova.compute.manager [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Instance network_info: |[{"id": "404c104c-258f-480e-bfe2-4ace4e83a43c", "address": "fa:16:3e:87:f8:2a", "network": {"id": "06017161-984c-4657-9639-4e6bdb6d4c09", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1072882119-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d45b5346eab44707b4d81eaaf12b0935", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39cd75b0-9ec7-48ed-b57f-34da0c573a60", "external-id": "nsx-vlan-transportzone-751", "segmentation_id": 751, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap404c104c-25", "ovs_interfaceid": "404c104c-258f-480e-bfe2-4ace4e83a43c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 816.501447] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:87:f8:2a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '39cd75b0-9ec7-48ed-b57f-34da0c573a60', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '404c104c-258f-480e-bfe2-4ace4e83a43c', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 816.508667] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Creating folder: Project (d45b5346eab44707b4d81eaaf12b0935). Parent ref: group-v261107. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 816.509474] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e257e98a-16be-4544-b4a9-b67a8273b870 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.524541] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Created folder: Project (d45b5346eab44707b4d81eaaf12b0935) in parent group-v261107. [ 816.524541] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Creating folder: Instances. Parent ref: group-v261202. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 816.524541] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bc3e0f5b-e672-404a-b103-8d86d948d294 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.542034] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Created folder: Instances in parent group-v261202. [ 816.543729] env[62385]: DEBUG oslo.service.loopingcall [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 816.543729] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 816.543729] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1813f05c-c286-4a08-8dc0-a33870e96d27 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.564298] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 816.564298] env[62385]: value = "task-1205801" [ 816.564298] env[62385]: _type = "Task" [ 816.564298] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.573819] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205801, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.651034] env[62385]: DEBUG oslo_vmware.api [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205798, 'name': ReconfigVM_Task, 'duration_secs': 0.359128} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.651300] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Reconfigured VM instance instance-00000042 to attach disk [datastore1] f2be792c-a723-4a69-b255-bbe7f0e8692d/f2be792c-a723-4a69-b255-bbe7f0e8692d.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 816.651945] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-95c7af8b-d2e8-426e-b47a-525af186bcc0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.659801] env[62385]: DEBUG oslo_vmware.api [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Waiting for the task: (returnval){ [ 816.659801] env[62385]: value = "task-1205802" [ 816.659801] env[62385]: _type = "Task" [ 816.659801] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.668946] env[62385]: DEBUG oslo_vmware.api [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205802, 'name': Rename_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.683877] env[62385]: DEBUG oslo_vmware.api [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52114399-8e43-161d-2d5a-b4f0990c647b, 'name': SearchDatastore_Task, 'duration_secs': 0.009594} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.684908] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b589428-816a-42f5-a7f9-773351c51ce1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.691558] env[62385]: DEBUG oslo_vmware.api [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 816.691558] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]522291c4-150c-d710-2e75-8a038b88b50d" [ 816.691558] env[62385]: _type = "Task" [ 816.691558] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.701524] env[62385]: DEBUG oslo_vmware.api [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]522291c4-150c-d710-2e75-8a038b88b50d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.983056] env[62385]: INFO nova.compute.manager [-] [instance: 2b744879-01d2-440e-8adf-58438b922888] Took 1.64 seconds to deallocate network for instance. [ 817.077753] env[62385]: DEBUG nova.compute.manager [req-f84e1ab1-a66c-428c-9770-7f9d62e10e2a req-1f49c396-c073-4f03-bcd0-f4a90f7fe670 service nova] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Received event network-changed-404c104c-258f-480e-bfe2-4ace4e83a43c {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 817.077962] env[62385]: DEBUG nova.compute.manager [req-f84e1ab1-a66c-428c-9770-7f9d62e10e2a req-1f49c396-c073-4f03-bcd0-f4a90f7fe670 service nova] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Refreshing instance network info cache due to event network-changed-404c104c-258f-480e-bfe2-4ace4e83a43c. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 817.078224] env[62385]: DEBUG oslo_concurrency.lockutils [req-f84e1ab1-a66c-428c-9770-7f9d62e10e2a req-1f49c396-c073-4f03-bcd0-f4a90f7fe670 service nova] Acquiring lock "refresh_cache-9b9fd167-d843-4b9d-9e4f-dbde5d3d629e" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.078444] env[62385]: DEBUG oslo_concurrency.lockutils [req-f84e1ab1-a66c-428c-9770-7f9d62e10e2a req-1f49c396-c073-4f03-bcd0-f4a90f7fe670 service nova] Acquired lock "refresh_cache-9b9fd167-d843-4b9d-9e4f-dbde5d3d629e" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.078665] env[62385]: DEBUG nova.network.neutron [req-f84e1ab1-a66c-428c-9770-7f9d62e10e2a req-1f49c396-c073-4f03-bcd0-f4a90f7fe670 service nova] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Refreshing network info cache for port 404c104c-258f-480e-bfe2-4ace4e83a43c {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 817.083991] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205801, 'name': CreateVM_Task, 'duration_secs': 0.471678} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.083991] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 817.085051] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.085051] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.085640] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 817.086956] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c3304d71-826d-49c9-b980-98faa360254b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.092996] env[62385]: DEBUG oslo_vmware.api [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 817.092996] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52232137-a2b3-cea8-ac55-95b7c82af6c0" [ 817.092996] env[62385]: _type = "Task" [ 817.092996] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.107141] env[62385]: DEBUG oslo_vmware.api [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52232137-a2b3-cea8-ac55-95b7c82af6c0, 'name': SearchDatastore_Task, 'duration_secs': 0.011019} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.107540] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.107680] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 817.107893] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.171606] env[62385]: DEBUG oslo_vmware.api [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205802, 'name': Rename_Task, 'duration_secs': 0.151469} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.174952] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 817.175568] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bb39d2fd-4e58-4f81-99d6-a2617bb4d738 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.184510] env[62385]: DEBUG oslo_vmware.api [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Waiting for the task: (returnval){ [ 817.184510] env[62385]: value = "task-1205803" [ 817.184510] env[62385]: _type = "Task" [ 817.184510] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.199482] env[62385]: DEBUG oslo_vmware.api [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205803, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.206164] env[62385]: DEBUG oslo_vmware.api [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]522291c4-150c-d710-2e75-8a038b88b50d, 'name': SearchDatastore_Task, 'duration_secs': 0.011571} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.209386] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.209988] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 55d62bba-ff30-42bf-b1cb-7567988a9361/55d62bba-ff30-42bf-b1cb-7567988a9361.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 817.210907] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.211382] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 817.212058] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fdedee67-1705-4747-bf90-6925b012a74f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.214573] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7cfe5410-8e0f-4a92-a0f9-d95af2e230f4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.226787] env[62385]: DEBUG oslo_vmware.api [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 817.226787] env[62385]: value = "task-1205804" [ 817.226787] env[62385]: _type = "Task" [ 817.226787] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.229798] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 817.230179] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 817.234734] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c1a662e-0d20-4d32-b373-6e86d55aa5fb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.253812] env[62385]: DEBUG oslo_vmware.api [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205804, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.254537] env[62385]: DEBUG oslo_vmware.api [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 817.254537] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52116d1f-c584-22cf-852e-ea9af7038430" [ 817.254537] env[62385]: _type = "Task" [ 817.254537] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.267999] env[62385]: DEBUG oslo_vmware.api [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52116d1f-c584-22cf-852e-ea9af7038430, 'name': SearchDatastore_Task, 'duration_secs': 0.010752} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.268819] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e16547b-2923-402c-8357-b7bae3f4aef7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.279113] env[62385]: DEBUG oslo_vmware.api [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 817.279113] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52b72702-4c89-0986-16b0-a69c9ff7c3ef" [ 817.279113] env[62385]: _type = "Task" [ 817.279113] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.288629] env[62385]: DEBUG oslo_vmware.api [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52b72702-4c89-0986-16b0-a69c9ff7c3ef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.337279] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc80c558-b9e2-4249-82ef-aa8c0aaddcce {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.347896] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a4a9138-068c-4214-a99f-5952fab1af33 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.381111] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da31e73e-6c04-4908-8b28-c57d854dd95e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.391219] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e913dc06-76b0-4fcb-90e2-6af0916d5ec7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.407078] env[62385]: DEBUG nova.compute.provider_tree [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 817.496082] env[62385]: DEBUG oslo_concurrency.lockutils [None req-695d2e36-0194-441c-bb9f-b1d1d915536f tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.702667] env[62385]: DEBUG oslo_vmware.api [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205803, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.741172] env[62385]: DEBUG oslo_vmware.api [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205804, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.793075] env[62385]: DEBUG oslo_vmware.api [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52b72702-4c89-0986-16b0-a69c9ff7c3ef, 'name': SearchDatastore_Task, 'duration_secs': 0.010037} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.793383] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.793782] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e/9b9fd167-d843-4b9d-9e4f-dbde5d3d629e.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 817.794281] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dd819711-08e7-40f7-b781-537b638d557f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.804379] env[62385]: DEBUG oslo_vmware.api [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 817.804379] env[62385]: value = "task-1205805" [ 817.804379] env[62385]: _type = "Task" [ 817.804379] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.814898] env[62385]: DEBUG oslo_vmware.api [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1205805, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.913406] env[62385]: DEBUG nova.scheduler.client.report [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 817.949882] env[62385]: DEBUG nova.network.neutron [req-f84e1ab1-a66c-428c-9770-7f9d62e10e2a req-1f49c396-c073-4f03-bcd0-f4a90f7fe670 service nova] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Updated VIF entry in instance network info cache for port 404c104c-258f-480e-bfe2-4ace4e83a43c. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 817.950265] env[62385]: DEBUG nova.network.neutron [req-f84e1ab1-a66c-428c-9770-7f9d62e10e2a req-1f49c396-c073-4f03-bcd0-f4a90f7fe670 service nova] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Updating instance_info_cache with network_info: [{"id": "404c104c-258f-480e-bfe2-4ace4e83a43c", "address": "fa:16:3e:87:f8:2a", "network": {"id": "06017161-984c-4657-9639-4e6bdb6d4c09", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1072882119-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d45b5346eab44707b4d81eaaf12b0935", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39cd75b0-9ec7-48ed-b57f-34da0c573a60", "external-id": "nsx-vlan-transportzone-751", "segmentation_id": 751, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap404c104c-25", "ovs_interfaceid": "404c104c-258f-480e-bfe2-4ace4e83a43c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.199862] env[62385]: DEBUG oslo_vmware.api [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205803, 'name': PowerOnVM_Task, 'duration_secs': 0.65686} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.200310] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 818.200421] env[62385]: INFO nova.compute.manager [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Took 9.16 seconds to spawn the instance on the hypervisor. [ 818.200584] env[62385]: DEBUG nova.compute.manager [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 818.201461] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b47caef3-e8b9-450d-bbde-ab686bd9286a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.243612] env[62385]: DEBUG oslo_vmware.api [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205804, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.531156} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.244243] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 55d62bba-ff30-42bf-b1cb-7567988a9361/55d62bba-ff30-42bf-b1cb-7567988a9361.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 818.244243] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 818.244510] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-41627bef-11fd-49b1-92cf-d09bc069045e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.253065] env[62385]: DEBUG oslo_vmware.api [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 818.253065] env[62385]: value = "task-1205806" [ 818.253065] env[62385]: _type = "Task" [ 818.253065] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.262703] env[62385]: DEBUG oslo_vmware.api [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205806, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.315244] env[62385]: DEBUG oslo_vmware.api [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1205805, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.475255} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.315518] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e/9b9fd167-d843-4b9d-9e4f-dbde5d3d629e.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 818.315740] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 818.316027] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-41b4fb61-8609-4e21-a0b9-5e11a84e4bd5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.324482] env[62385]: DEBUG oslo_vmware.api [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 818.324482] env[62385]: value = "task-1205807" [ 818.324482] env[62385]: _type = "Task" [ 818.324482] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.334376] env[62385]: DEBUG oslo_vmware.api [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1205807, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.431255] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.497s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.431255] env[62385]: DEBUG nova.compute.manager [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 818.434789] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ebabc504-fb40-4408-abaf-177b5b541c69 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.195s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.435203] env[62385]: DEBUG nova.objects.instance [None req-ebabc504-fb40-4408-abaf-177b5b541c69 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Lazy-loading 'resources' on Instance uuid f50ddafe-f5a9-422b-b0f4-46c1b111dfbe {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 818.453065] env[62385]: DEBUG oslo_concurrency.lockutils [req-f84e1ab1-a66c-428c-9770-7f9d62e10e2a req-1f49c396-c073-4f03-bcd0-f4a90f7fe670 service nova] Releasing lock "refresh_cache-9b9fd167-d843-4b9d-9e4f-dbde5d3d629e" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.453325] env[62385]: DEBUG nova.compute.manager [req-f84e1ab1-a66c-428c-9770-7f9d62e10e2a req-1f49c396-c073-4f03-bcd0-f4a90f7fe670 service nova] [instance: 2b744879-01d2-440e-8adf-58438b922888] Received event network-vif-deleted-f42dfa8b-5ef4-40f7-9359-aa425ae2fbad {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 818.732589] env[62385]: INFO nova.compute.manager [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Took 42.22 seconds to build instance. [ 818.765854] env[62385]: DEBUG oslo_vmware.api [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205806, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.17196} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.766242] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 818.767414] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7c7dd51-da4a-4093-8662-7c31003c2521 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.797878] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] 55d62bba-ff30-42bf-b1cb-7567988a9361/55d62bba-ff30-42bf-b1cb-7567988a9361.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 818.798268] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9cdcc6b4-fca7-4b7c-89c3-cec34284db9b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.820429] env[62385]: DEBUG oslo_vmware.api [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 818.820429] env[62385]: value = "task-1205808" [ 818.820429] env[62385]: _type = "Task" [ 818.820429] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.831442] env[62385]: DEBUG oslo_vmware.api [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205808, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.836893] env[62385]: DEBUG oslo_vmware.api [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1205807, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.3508} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.837177] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 818.837960] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e63697c1-7017-4fea-b4e0-15fcd81ad99f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.864670] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e/9b9fd167-d843-4b9d-9e4f-dbde5d3d629e.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 818.865171] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f1e73c9f-65ed-4554-8bc4-486f96708add {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.888304] env[62385]: DEBUG oslo_vmware.api [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 818.888304] env[62385]: value = "task-1205809" [ 818.888304] env[62385]: _type = "Task" [ 818.888304] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.896594] env[62385]: DEBUG oslo_vmware.api [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1205809, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.936880] env[62385]: DEBUG nova.compute.utils [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 818.941021] env[62385]: DEBUG nova.compute.manager [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 818.941327] env[62385]: DEBUG nova.network.neutron [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 819.009385] env[62385]: DEBUG nova.policy [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '825ffbadf064473aa5eda331914306f2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7ac0c8638f064daba042ded4de8c1cc5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 819.234724] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d1f7103a-52b8-411f-a16a-46fbff01a05c tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Lock "f2be792c-a723-4a69-b255-bbe7f0e8692d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 91.845s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.326737] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3d826b4-05f1-4804-9a65-58ec337a64ec {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.334974] env[62385]: DEBUG oslo_vmware.api [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205808, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.340179] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7219005-2240-44b0-b586-41aac13a3103 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.376938] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9558415e-7afb-4f94-889e-3325af274f6a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.398223] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ea6db64-72b9-4ba0-8d18-9c2c5fa75b97 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.404448] env[62385]: DEBUG oslo_vmware.api [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1205809, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.412565] env[62385]: DEBUG nova.compute.provider_tree [None req-ebabc504-fb40-4408-abaf-177b5b541c69 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 819.443715] env[62385]: DEBUG nova.compute.manager [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 819.453117] env[62385]: DEBUG nova.network.neutron [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Successfully created port: f5b51ed4-e913-4534-a2d6-33cf95ab3041 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 819.833363] env[62385]: DEBUG oslo_vmware.api [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205808, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.901799] env[62385]: DEBUG oslo_vmware.api [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1205809, 'name': ReconfigVM_Task, 'duration_secs': 0.990247} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.902607] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Reconfigured VM instance instance-00000044 to attach disk [datastore1] 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e/9b9fd167-d843-4b9d-9e4f-dbde5d3d629e.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 819.902868] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-202c7897-168b-4889-a986-c6ecfeac6389 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.910145] env[62385]: DEBUG oslo_vmware.api [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 819.910145] env[62385]: value = "task-1205810" [ 819.910145] env[62385]: _type = "Task" [ 819.910145] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.919900] env[62385]: DEBUG nova.scheduler.client.report [None req-ebabc504-fb40-4408-abaf-177b5b541c69 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 819.923288] env[62385]: DEBUG oslo_vmware.api [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1205810, 'name': Rename_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.336962] env[62385]: DEBUG oslo_vmware.api [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205808, 'name': ReconfigVM_Task, 'duration_secs': 1.027265} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.337656] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Reconfigured VM instance instance-00000043 to attach disk [datastore1] 55d62bba-ff30-42bf-b1cb-7567988a9361/55d62bba-ff30-42bf-b1cb-7567988a9361.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 820.338582] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-813d155f-c957-4c87-bf10-412d55c9cd59 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.347245] env[62385]: DEBUG oslo_vmware.api [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 820.347245] env[62385]: value = "task-1205811" [ 820.347245] env[62385]: _type = "Task" [ 820.347245] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.355426] env[62385]: DEBUG oslo_vmware.api [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205811, 'name': Rename_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.421189] env[62385]: DEBUG oslo_vmware.api [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1205810, 'name': Rename_Task, 'duration_secs': 0.263389} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.421584] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 820.421876] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b7192364-19fb-4d16-9d6f-2502d4bb2220 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.425248] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ebabc504-fb40-4408-abaf-177b5b541c69 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.990s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.429169] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 31.729s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.429169] env[62385]: DEBUG nova.objects.instance [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62385) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 820.432219] env[62385]: DEBUG oslo_vmware.api [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 820.432219] env[62385]: value = "task-1205812" [ 820.432219] env[62385]: _type = "Task" [ 820.432219] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.444809] env[62385]: DEBUG oslo_vmware.api [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1205812, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.457027] env[62385]: DEBUG nova.compute.manager [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 820.458953] env[62385]: INFO nova.scheduler.client.report [None req-ebabc504-fb40-4408-abaf-177b5b541c69 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Deleted allocations for instance f50ddafe-f5a9-422b-b0f4-46c1b111dfbe [ 820.491618] env[62385]: DEBUG nova.virt.hardware [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 820.492093] env[62385]: DEBUG nova.virt.hardware [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 820.492256] env[62385]: DEBUG nova.virt.hardware [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 820.492458] env[62385]: DEBUG nova.virt.hardware [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 820.496035] env[62385]: DEBUG nova.virt.hardware [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 820.496035] env[62385]: DEBUG nova.virt.hardware [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 820.496035] env[62385]: DEBUG nova.virt.hardware [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 820.496035] env[62385]: DEBUG nova.virt.hardware [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 820.496035] env[62385]: DEBUG nova.virt.hardware [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 820.496403] env[62385]: DEBUG nova.virt.hardware [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 820.496403] env[62385]: DEBUG nova.virt.hardware [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 820.496403] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86ca7173-7d1d-46e2-aafe-58653d9be227 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.505616] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f021f99a-a0d4-4e27-ae6f-3410c08d3ffc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.856869] env[62385]: DEBUG oslo_vmware.api [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205811, 'name': Rename_Task, 'duration_secs': 0.188148} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.857191] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 820.857454] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8d7285df-50b9-40fa-917c-2ebc3e97c7bd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.865958] env[62385]: DEBUG oslo_vmware.api [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 820.865958] env[62385]: value = "task-1205813" [ 820.865958] env[62385]: _type = "Task" [ 820.865958] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.876789] env[62385]: DEBUG oslo_vmware.api [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205813, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.949292] env[62385]: DEBUG nova.compute.manager [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 820.949809] env[62385]: DEBUG oslo_vmware.api [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1205812, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.951048] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8881920-1c8a-4ec3-ad00-139627056517 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.970682] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ebabc504-fb40-4408-abaf-177b5b541c69 tempest-AttachInterfacesUnderV243Test-187544623 tempest-AttachInterfacesUnderV243Test-187544623-project-member] Lock "f50ddafe-f5a9-422b-b0f4-46c1b111dfbe" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.120s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.015891] env[62385]: DEBUG nova.compute.manager [req-6a3d7fcd-d680-4de4-8ad5-8ba58d0f5a58 req-a27a21a6-595b-4d47-a8e9-1249d567ac6c service nova] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Received event network-vif-plugged-f5b51ed4-e913-4534-a2d6-33cf95ab3041 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 821.016140] env[62385]: DEBUG oslo_concurrency.lockutils [req-6a3d7fcd-d680-4de4-8ad5-8ba58d0f5a58 req-a27a21a6-595b-4d47-a8e9-1249d567ac6c service nova] Acquiring lock "679150f5-ad17-45e6-9e3d-9f6e28b1aee3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.016408] env[62385]: DEBUG oslo_concurrency.lockutils [req-6a3d7fcd-d680-4de4-8ad5-8ba58d0f5a58 req-a27a21a6-595b-4d47-a8e9-1249d567ac6c service nova] Lock "679150f5-ad17-45e6-9e3d-9f6e28b1aee3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.016588] env[62385]: DEBUG oslo_concurrency.lockutils [req-6a3d7fcd-d680-4de4-8ad5-8ba58d0f5a58 req-a27a21a6-595b-4d47-a8e9-1249d567ac6c service nova] Lock "679150f5-ad17-45e6-9e3d-9f6e28b1aee3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.016699] env[62385]: DEBUG nova.compute.manager [req-6a3d7fcd-d680-4de4-8ad5-8ba58d0f5a58 req-a27a21a6-595b-4d47-a8e9-1249d567ac6c service nova] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] No waiting events found dispatching network-vif-plugged-f5b51ed4-e913-4534-a2d6-33cf95ab3041 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 821.017024] env[62385]: WARNING nova.compute.manager [req-6a3d7fcd-d680-4de4-8ad5-8ba58d0f5a58 req-a27a21a6-595b-4d47-a8e9-1249d567ac6c service nova] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Received unexpected event network-vif-plugged-f5b51ed4-e913-4534-a2d6-33cf95ab3041 for instance with vm_state building and task_state spawning. [ 821.061493] env[62385]: DEBUG nova.network.neutron [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Successfully updated port: f5b51ed4-e913-4534-a2d6-33cf95ab3041 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 821.378628] env[62385]: DEBUG oslo_vmware.api [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205813, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.442223] env[62385]: DEBUG oslo_vmware.api [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1205812, 'name': PowerOnVM_Task, 'duration_secs': 0.564063} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.442490] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 821.442734] env[62385]: INFO nova.compute.manager [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Took 7.33 seconds to spawn the instance on the hypervisor. [ 821.442934] env[62385]: DEBUG nova.compute.manager [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 821.443725] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a51b9f9c-c777-4d85-9b3e-8cb744f81ae0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.447128] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fc29e912-581e-4c84-bec8-b1b7c2eda53f tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.020s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.448466] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.923s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.450030] env[62385]: INFO nova.compute.claims [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 821.467080] env[62385]: INFO nova.compute.manager [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] instance snapshotting [ 821.469838] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6f754cf-df74-4c7f-a87d-15075dbc4b98 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.494279] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4452c92-8e40-4eea-862d-2b88f3a0d2b2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.565747] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquiring lock "refresh_cache-679150f5-ad17-45e6-9e3d-9f6e28b1aee3" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.565918] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquired lock "refresh_cache-679150f5-ad17-45e6-9e3d-9f6e28b1aee3" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.566101] env[62385]: DEBUG nova.network.neutron [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 821.880029] env[62385]: DEBUG oslo_vmware.api [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205813, 'name': PowerOnVM_Task, 'duration_secs': 0.998051} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.880539] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 821.880756] env[62385]: INFO nova.compute.manager [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Took 10.29 seconds to spawn the instance on the hypervisor. [ 821.880942] env[62385]: DEBUG nova.compute.manager [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 821.881916] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c45fc27-4fed-461c-b564-38a53ef445dd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.969524] env[62385]: INFO nova.compute.manager [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Took 42.08 seconds to build instance. [ 822.006725] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Creating Snapshot of the VM instance {{(pid=62385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 822.007050] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-9a1d903e-59bb-4121-8c58-36f94047d508 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.017788] env[62385]: DEBUG oslo_vmware.api [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Waiting for the task: (returnval){ [ 822.017788] env[62385]: value = "task-1205814" [ 822.017788] env[62385]: _type = "Task" [ 822.017788] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.029461] env[62385]: DEBUG oslo_vmware.api [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205814, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.128494] env[62385]: DEBUG nova.network.neutron [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 822.365134] env[62385]: DEBUG nova.network.neutron [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Updating instance_info_cache with network_info: [{"id": "f5b51ed4-e913-4534-a2d6-33cf95ab3041", "address": "fa:16:3e:7f:d8:08", "network": {"id": "b0a3dea4-1925-4920-8a8e-9dc66733d1da", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1951021425-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ac0c8638f064daba042ded4de8c1cc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee20e439-fed9-490e-97dd-f3c886977ae1", "external-id": "nsx-vlan-transportzone-357", "segmentation_id": 357, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5b51ed4-e9", "ovs_interfaceid": "f5b51ed4-e913-4534-a2d6-33cf95ab3041", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.404711] env[62385]: INFO nova.compute.manager [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Took 43.07 seconds to build instance. [ 822.472049] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f93e6a82-1698-4ce2-bb8f-90bbbc971a11 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "9b9fd167-d843-4b9d-9e4f-dbde5d3d629e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 91.131s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.529723] env[62385]: DEBUG oslo_vmware.api [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205814, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.851593] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b17e4649-4602-4971-980a-7f8d6c864c2a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.862396] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09f9a000-88eb-4361-92dd-6482f43e7d19 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.868048] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Releasing lock "refresh_cache-679150f5-ad17-45e6-9e3d-9f6e28b1aee3" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.868369] env[62385]: DEBUG nova.compute.manager [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Instance network_info: |[{"id": "f5b51ed4-e913-4534-a2d6-33cf95ab3041", "address": "fa:16:3e:7f:d8:08", "network": {"id": "b0a3dea4-1925-4920-8a8e-9dc66733d1da", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1951021425-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ac0c8638f064daba042ded4de8c1cc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee20e439-fed9-490e-97dd-f3c886977ae1", "external-id": "nsx-vlan-transportzone-357", "segmentation_id": 357, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5b51ed4-e9", "ovs_interfaceid": "f5b51ed4-e913-4534-a2d6-33cf95ab3041", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 822.868851] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7f:d8:08', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ee20e439-fed9-490e-97dd-f3c886977ae1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f5b51ed4-e913-4534-a2d6-33cf95ab3041', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 822.877226] env[62385]: DEBUG oslo.service.loopingcall [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 822.908491] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 822.908491] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8f7ce39d-2b7c-4fd4-bff8-278ede3c01d0 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "55d62bba-ff30-42bf-b1cb-7567988a9361" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 94.245s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.908766] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3252ad53-882b-40d8-8342-4cc2df026d8d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.928246] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a134699d-1532-4340-8998-4b935584fbea {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.938977] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c3c8744-8a5a-4d21-8864-7fd6114f2ec3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.945035] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 822.945035] env[62385]: value = "task-1205815" [ 822.945035] env[62385]: _type = "Task" [ 822.945035] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.954931] env[62385]: DEBUG nova.compute.provider_tree [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 822.962726] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205815, 'name': CreateVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.029374] env[62385]: DEBUG oslo_vmware.api [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205814, 'name': CreateSnapshot_Task, 'duration_secs': 0.859577} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.029732] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Created Snapshot of the VM instance {{(pid=62385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 823.030667] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5ee9984-5430-4831-86ff-bd35ec8fd856 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.049177] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1980d67d-7399-48ce-85a3-b137ec931d83 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.057054] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a77e36e5-4e29-4f89-afcb-513e93f7ae76 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Suspending the VM {{(pid=62385) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 823.058252] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-75277a2a-e8a8-4b30-86a4-4f9026fd162d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.062579] env[62385]: DEBUG nova.compute.manager [req-be5f11ee-9484-43ee-9b8b-1f42de30a341 req-d5b2c5bd-7538-45cf-9ea8-056009a5714e service nova] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Received event network-changed-f5b51ed4-e913-4534-a2d6-33cf95ab3041 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 823.062814] env[62385]: DEBUG nova.compute.manager [req-be5f11ee-9484-43ee-9b8b-1f42de30a341 req-d5b2c5bd-7538-45cf-9ea8-056009a5714e service nova] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Refreshing instance network info cache due to event network-changed-f5b51ed4-e913-4534-a2d6-33cf95ab3041. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 823.063092] env[62385]: DEBUG oslo_concurrency.lockutils [req-be5f11ee-9484-43ee-9b8b-1f42de30a341 req-d5b2c5bd-7538-45cf-9ea8-056009a5714e service nova] Acquiring lock "refresh_cache-679150f5-ad17-45e6-9e3d-9f6e28b1aee3" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.063279] env[62385]: DEBUG oslo_concurrency.lockutils [req-be5f11ee-9484-43ee-9b8b-1f42de30a341 req-d5b2c5bd-7538-45cf-9ea8-056009a5714e service nova] Acquired lock "refresh_cache-679150f5-ad17-45e6-9e3d-9f6e28b1aee3" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.063475] env[62385]: DEBUG nova.network.neutron [req-be5f11ee-9484-43ee-9b8b-1f42de30a341 req-d5b2c5bd-7538-45cf-9ea8-056009a5714e service nova] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Refreshing network info cache for port f5b51ed4-e913-4534-a2d6-33cf95ab3041 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 823.071665] env[62385]: DEBUG oslo_vmware.api [None req-a77e36e5-4e29-4f89-afcb-513e93f7ae76 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 823.071665] env[62385]: value = "task-1205816" [ 823.071665] env[62385]: _type = "Task" [ 823.071665] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.084675] env[62385]: DEBUG oslo_vmware.api [None req-a77e36e5-4e29-4f89-afcb-513e93f7ae76 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205816, 'name': SuspendVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.454834] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205815, 'name': CreateVM_Task, 'duration_secs': 0.380691} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.455207] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 823.455732] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.455904] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.456259] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 823.456631] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-556439f5-0f04-46da-8218-f98f9e1a90a8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.459112] env[62385]: DEBUG nova.scheduler.client.report [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 823.468063] env[62385]: DEBUG oslo_vmware.api [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 823.468063] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5216b43c-5c28-19bf-28e8-4f1353ff0942" [ 823.468063] env[62385]: _type = "Task" [ 823.468063] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.482700] env[62385]: DEBUG oslo_vmware.api [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5216b43c-5c28-19bf-28e8-4f1353ff0942, 'name': SearchDatastore_Task, 'duration_secs': 0.011384} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.483558] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.483807] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 823.484058] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.484254] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.485434] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 823.485434] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f1a1e258-9a48-4f57-8b96-39af8a2462af {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.495386] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 823.495822] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 823.496381] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fb7547b3-99dd-4e1f-b3a2-143c6a9d0435 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.504302] env[62385]: DEBUG oslo_vmware.api [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 823.504302] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521436bf-49ca-2ca0-7fa5-80962401a1f0" [ 823.504302] env[62385]: _type = "Task" [ 823.504302] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.513859] env[62385]: DEBUG oslo_vmware.api [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521436bf-49ca-2ca0-7fa5-80962401a1f0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.562251] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Creating linked-clone VM from snapshot {{(pid=62385) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 823.562850] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-47ec6ef1-3304-49ff-93c3-9acdd69e7e57 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.581731] env[62385]: DEBUG oslo_vmware.api [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Waiting for the task: (returnval){ [ 823.581731] env[62385]: value = "task-1205817" [ 823.581731] env[62385]: _type = "Task" [ 823.581731] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.591814] env[62385]: DEBUG oslo_vmware.api [None req-a77e36e5-4e29-4f89-afcb-513e93f7ae76 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205816, 'name': SuspendVM_Task} progress is 62%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.601031] env[62385]: DEBUG oslo_vmware.api [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205817, 'name': CloneVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.827815] env[62385]: DEBUG nova.network.neutron [req-be5f11ee-9484-43ee-9b8b-1f42de30a341 req-d5b2c5bd-7538-45cf-9ea8-056009a5714e service nova] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Updated VIF entry in instance network info cache for port f5b51ed4-e913-4534-a2d6-33cf95ab3041. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 823.827815] env[62385]: DEBUG nova.network.neutron [req-be5f11ee-9484-43ee-9b8b-1f42de30a341 req-d5b2c5bd-7538-45cf-9ea8-056009a5714e service nova] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Updating instance_info_cache with network_info: [{"id": "f5b51ed4-e913-4534-a2d6-33cf95ab3041", "address": "fa:16:3e:7f:d8:08", "network": {"id": "b0a3dea4-1925-4920-8a8e-9dc66733d1da", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1951021425-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ac0c8638f064daba042ded4de8c1cc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee20e439-fed9-490e-97dd-f3c886977ae1", "external-id": "nsx-vlan-transportzone-357", "segmentation_id": 357, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5b51ed4-e9", "ovs_interfaceid": "f5b51ed4-e913-4534-a2d6-33cf95ab3041", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.967721] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.519s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.968309] env[62385]: DEBUG nova.compute.manager [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 823.971085] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.831s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.971326] env[62385]: DEBUG nova.objects.instance [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lazy-loading 'resources' on Instance uuid f8a98710-9d02-4aa3-b6a5-34b7060ad62d {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 824.015966] env[62385]: DEBUG oslo_vmware.api [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521436bf-49ca-2ca0-7fa5-80962401a1f0, 'name': SearchDatastore_Task, 'duration_secs': 0.009315} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.016899] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cdc1f5d4-23ba-487b-a4eb-3e80bf40f124 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.024649] env[62385]: DEBUG oslo_vmware.api [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 824.024649] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52b5ec1a-1b70-8379-96c5-ca3ff55a7b7d" [ 824.024649] env[62385]: _type = "Task" [ 824.024649] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.034811] env[62385]: DEBUG oslo_vmware.api [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52b5ec1a-1b70-8379-96c5-ca3ff55a7b7d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.085929] env[62385]: DEBUG oslo_vmware.api [None req-a77e36e5-4e29-4f89-afcb-513e93f7ae76 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205816, 'name': SuspendVM_Task, 'duration_secs': 0.856784} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.090644] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a77e36e5-4e29-4f89-afcb-513e93f7ae76 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Suspended the VM {{(pid=62385) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 824.090930] env[62385]: DEBUG nova.compute.manager [None req-a77e36e5-4e29-4f89-afcb-513e93f7ae76 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 824.091978] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e6f49a5-4ea4-43dd-8f85-8834e991aa88 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.105126] env[62385]: DEBUG oslo_vmware.api [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205817, 'name': CloneVM_Task} progress is 94%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.117126] env[62385]: DEBUG nova.compute.manager [req-d248ae0c-855d-4fec-9d86-6936c4eca06d req-e114a215-c100-460c-a5bf-72327f6876b7 service nova] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Received event network-changed-404c104c-258f-480e-bfe2-4ace4e83a43c {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 824.117452] env[62385]: DEBUG nova.compute.manager [req-d248ae0c-855d-4fec-9d86-6936c4eca06d req-e114a215-c100-460c-a5bf-72327f6876b7 service nova] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Refreshing instance network info cache due to event network-changed-404c104c-258f-480e-bfe2-4ace4e83a43c. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 824.117774] env[62385]: DEBUG oslo_concurrency.lockutils [req-d248ae0c-855d-4fec-9d86-6936c4eca06d req-e114a215-c100-460c-a5bf-72327f6876b7 service nova] Acquiring lock "refresh_cache-9b9fd167-d843-4b9d-9e4f-dbde5d3d629e" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.117966] env[62385]: DEBUG oslo_concurrency.lockutils [req-d248ae0c-855d-4fec-9d86-6936c4eca06d req-e114a215-c100-460c-a5bf-72327f6876b7 service nova] Acquired lock "refresh_cache-9b9fd167-d843-4b9d-9e4f-dbde5d3d629e" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.118189] env[62385]: DEBUG nova.network.neutron [req-d248ae0c-855d-4fec-9d86-6936c4eca06d req-e114a215-c100-460c-a5bf-72327f6876b7 service nova] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Refreshing network info cache for port 404c104c-258f-480e-bfe2-4ace4e83a43c {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 824.330769] env[62385]: DEBUG oslo_concurrency.lockutils [req-be5f11ee-9484-43ee-9b8b-1f42de30a341 req-d5b2c5bd-7538-45cf-9ea8-056009a5714e service nova] Releasing lock "refresh_cache-679150f5-ad17-45e6-9e3d-9f6e28b1aee3" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.474381] env[62385]: DEBUG nova.objects.instance [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lazy-loading 'numa_topology' on Instance uuid f8a98710-9d02-4aa3-b6a5-34b7060ad62d {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 824.476776] env[62385]: DEBUG nova.compute.utils [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 824.481542] env[62385]: DEBUG nova.compute.manager [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 824.481720] env[62385]: DEBUG nova.network.neutron [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 824.534919] env[62385]: DEBUG oslo_vmware.api [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52b5ec1a-1b70-8379-96c5-ca3ff55a7b7d, 'name': SearchDatastore_Task, 'duration_secs': 0.010339} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.536415] env[62385]: DEBUG nova.policy [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4ac552dc01fc4de2b97e36b41dae725a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '105150c170d74e268e82ab894e9bf0d5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 824.537848] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.538786] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 679150f5-ad17-45e6-9e3d-9f6e28b1aee3/679150f5-ad17-45e6-9e3d-9f6e28b1aee3.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 824.538786] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-17b7a3d4-d5d5-48a5-8d34-129c290f9216 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.546943] env[62385]: DEBUG oslo_vmware.api [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 824.546943] env[62385]: value = "task-1205818" [ 824.546943] env[62385]: _type = "Task" [ 824.546943] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.556012] env[62385]: DEBUG oslo_vmware.api [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205818, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.599317] env[62385]: DEBUG oslo_vmware.api [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205817, 'name': CloneVM_Task} progress is 94%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.853282] env[62385]: DEBUG nova.network.neutron [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Successfully created port: 844976e8-2627-499f-9b01-284d376c86fa {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 824.983226] env[62385]: DEBUG nova.objects.base [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=62385) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 824.988277] env[62385]: DEBUG nova.compute.manager [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 825.064349] env[62385]: DEBUG oslo_vmware.api [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205818, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.501642} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.065236] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 679150f5-ad17-45e6-9e3d-9f6e28b1aee3/679150f5-ad17-45e6-9e3d-9f6e28b1aee3.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 825.065427] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 825.065739] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-aa51b286-5cc4-4092-aa52-f6a9f8417b1c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.079035] env[62385]: DEBUG oslo_vmware.api [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 825.079035] env[62385]: value = "task-1205819" [ 825.079035] env[62385]: _type = "Task" [ 825.079035] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.087195] env[62385]: DEBUG oslo_vmware.api [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205819, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.099524] env[62385]: DEBUG oslo_vmware.api [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205817, 'name': CloneVM_Task, 'duration_secs': 1.520903} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.099524] env[62385]: INFO nova.virt.vmwareapi.vmops [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Created linked-clone VM from snapshot [ 825.101874] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e15f2f2-4149-44dc-bfb0-d97651893c3a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.111456] env[62385]: DEBUG nova.virt.vmwareapi.images [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Uploading image 4c74362b-3503-4ff3-8801-a50effea5d68 {{(pid=62385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 825.143590] env[62385]: DEBUG oslo_vmware.rw_handles [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 825.143590] env[62385]: value = "vm-261207" [ 825.143590] env[62385]: _type = "VirtualMachine" [ 825.143590] env[62385]: }. {{(pid=62385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 825.144996] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-17c1e08d-d9de-4b61-a19d-426535273b54 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.158657] env[62385]: DEBUG oslo_vmware.rw_handles [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Lease: (returnval){ [ 825.158657] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]520aeffc-2bd4-120e-96ba-c0bfdb29d653" [ 825.158657] env[62385]: _type = "HttpNfcLease" [ 825.158657] env[62385]: } obtained for exporting VM: (result){ [ 825.158657] env[62385]: value = "vm-261207" [ 825.158657] env[62385]: _type = "VirtualMachine" [ 825.158657] env[62385]: }. {{(pid=62385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 825.158657] env[62385]: DEBUG oslo_vmware.api [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Waiting for the lease: (returnval){ [ 825.158657] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]520aeffc-2bd4-120e-96ba-c0bfdb29d653" [ 825.158657] env[62385]: _type = "HttpNfcLease" [ 825.158657] env[62385]: } to be ready. {{(pid=62385) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 825.167450] env[62385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 825.167450] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]520aeffc-2bd4-120e-96ba-c0bfdb29d653" [ 825.167450] env[62385]: _type = "HttpNfcLease" [ 825.167450] env[62385]: } is initializing. {{(pid=62385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 825.337718] env[62385]: DEBUG nova.network.neutron [req-d248ae0c-855d-4fec-9d86-6936c4eca06d req-e114a215-c100-460c-a5bf-72327f6876b7 service nova] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Updated VIF entry in instance network info cache for port 404c104c-258f-480e-bfe2-4ace4e83a43c. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 825.337825] env[62385]: DEBUG nova.network.neutron [req-d248ae0c-855d-4fec-9d86-6936c4eca06d req-e114a215-c100-460c-a5bf-72327f6876b7 service nova] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Updating instance_info_cache with network_info: [{"id": "404c104c-258f-480e-bfe2-4ace4e83a43c", "address": "fa:16:3e:87:f8:2a", "network": {"id": "06017161-984c-4657-9639-4e6bdb6d4c09", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1072882119-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.208", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d45b5346eab44707b4d81eaaf12b0935", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39cd75b0-9ec7-48ed-b57f-34da0c573a60", "external-id": "nsx-vlan-transportzone-751", "segmentation_id": 751, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap404c104c-25", "ovs_interfaceid": "404c104c-258f-480e-bfe2-4ace4e83a43c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.431220] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37125435-8221-46c3-98ac-815fe62330e4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.441168] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d4fc32e-9bb2-4f52-8ccf-b39f42d173c0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.486210] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-537328ce-ae0a-4141-83bb-42a6d5338867 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.493265] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbb3a7fc-f83b-4fd7-b8c2-25ac44e8fca8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.516454] env[62385]: DEBUG nova.compute.provider_tree [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Updating inventory in ProviderTree for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 825.586539] env[62385]: DEBUG oslo_vmware.api [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205819, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.286244} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.590019] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 825.590019] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0e52122-aefb-47bb-a4a8-5bff744d2b7c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.619093] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] 679150f5-ad17-45e6-9e3d-9f6e28b1aee3/679150f5-ad17-45e6-9e3d-9f6e28b1aee3.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 825.619943] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b9098ec2-94c6-4248-b152-8eecbcad1d29 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.645774] env[62385]: DEBUG oslo_vmware.api [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 825.645774] env[62385]: value = "task-1205821" [ 825.645774] env[62385]: _type = "Task" [ 825.645774] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.657191] env[62385]: DEBUG oslo_vmware.api [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205821, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.668281] env[62385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 825.668281] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]520aeffc-2bd4-120e-96ba-c0bfdb29d653" [ 825.668281] env[62385]: _type = "HttpNfcLease" [ 825.668281] env[62385]: } is ready. {{(pid=62385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 825.668281] env[62385]: DEBUG oslo_vmware.rw_handles [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 825.668281] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]520aeffc-2bd4-120e-96ba-c0bfdb29d653" [ 825.668281] env[62385]: _type = "HttpNfcLease" [ 825.668281] env[62385]: }. {{(pid=62385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 825.668717] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fe908fe-db04-4444-81cc-cb104a0f0b9a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.677642] env[62385]: DEBUG oslo_vmware.rw_handles [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b395f0-8a10-8613-b1bd-bd588ab77f51/disk-0.vmdk from lease info. {{(pid=62385) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 825.678460] env[62385]: DEBUG oslo_vmware.rw_handles [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b395f0-8a10-8613-b1bd-bd588ab77f51/disk-0.vmdk for reading. {{(pid=62385) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 825.820705] env[62385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-4da23833-6ea1-41fa-a8b9-6f1b6c47251e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.844728] env[62385]: DEBUG oslo_concurrency.lockutils [req-d248ae0c-855d-4fec-9d86-6936c4eca06d req-e114a215-c100-460c-a5bf-72327f6876b7 service nova] Releasing lock "refresh_cache-9b9fd167-d843-4b9d-9e4f-dbde5d3d629e" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.005686] env[62385]: DEBUG nova.compute.manager [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 826.048180] env[62385]: ERROR nova.scheduler.client.report [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [req-9b6f3cea-657b-466c-affe-625e0c6e6370] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-9b6f3cea-657b-466c-affe-625e0c6e6370"}]} [ 826.067043] env[62385]: DEBUG nova.virt.hardware [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 826.069547] env[62385]: DEBUG nova.virt.hardware [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 826.070019] env[62385]: DEBUG nova.virt.hardware [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 826.070148] env[62385]: DEBUG nova.virt.hardware [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 826.070392] env[62385]: DEBUG nova.virt.hardware [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 826.070844] env[62385]: DEBUG nova.virt.hardware [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 826.071110] env[62385]: DEBUG nova.virt.hardware [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 826.071283] env[62385]: DEBUG nova.virt.hardware [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 826.071470] env[62385]: DEBUG nova.virt.hardware [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 826.071650] env[62385]: DEBUG nova.virt.hardware [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 826.071876] env[62385]: DEBUG nova.virt.hardware [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 826.073025] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b2fb3b5-f9a1-416e-aeae-7cf07ad6b35e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.076650] env[62385]: DEBUG nova.scheduler.client.report [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Refreshing inventories for resource provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 826.085429] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26c527b8-4a56-4a3f-b352-9b2e2c33f2c3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.108399] env[62385]: DEBUG nova.scheduler.client.report [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Updating ProviderTree inventory for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 826.109273] env[62385]: DEBUG nova.compute.provider_tree [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Updating inventory in ProviderTree for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 826.121771] env[62385]: DEBUG nova.scheduler.client.report [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Refreshing aggregate associations for resource provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a, aggregates: None {{(pid=62385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 826.158036] env[62385]: DEBUG oslo_vmware.api [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205821, 'name': ReconfigVM_Task, 'duration_secs': 0.31576} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.159127] env[62385]: DEBUG nova.scheduler.client.report [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Refreshing trait associations for resource provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=62385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 826.161284] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Reconfigured VM instance instance-00000045 to attach disk [datastore2] 679150f5-ad17-45e6-9e3d-9f6e28b1aee3/679150f5-ad17-45e6-9e3d-9f6e28b1aee3.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 826.162261] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-052937e8-feaa-43e4-b3a4-2f3761e21b94 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.169832] env[62385]: DEBUG oslo_vmware.api [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 826.169832] env[62385]: value = "task-1205822" [ 826.169832] env[62385]: _type = "Task" [ 826.169832] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.180086] env[62385]: DEBUG oslo_vmware.api [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205822, 'name': Rename_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.273460] env[62385]: DEBUG oslo_concurrency.lockutils [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Acquiring lock "699a57a9-5a1a-4cd1-8449-723400d9caf4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.273769] env[62385]: DEBUG oslo_concurrency.lockutils [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Lock "699a57a9-5a1a-4cd1-8449-723400d9caf4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.553222] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57dcb965-26cf-4e8b-9257-2aef366dc24a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.562845] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5da0f34-b4df-4e26-abea-d1fdd14b94b7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.604615] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f70e25de-98a8-49b6-916d-09968515afd8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.613759] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1ce5c13-6141-42d6-a0c5-300db22c128a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.629479] env[62385]: DEBUG nova.compute.provider_tree [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Updating inventory in ProviderTree for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 826.683540] env[62385]: DEBUG oslo_vmware.api [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205822, 'name': Rename_Task, 'duration_secs': 0.168177} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.684090] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 826.684268] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-eaa7c31e-eb9c-4954-8c5e-967c612f4f11 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.697530] env[62385]: DEBUG oslo_vmware.api [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 826.697530] env[62385]: value = "task-1205823" [ 826.697530] env[62385]: _type = "Task" [ 826.697530] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.715759] env[62385]: DEBUG oslo_vmware.api [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205823, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.778171] env[62385]: DEBUG nova.compute.manager [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 826.878132] env[62385]: DEBUG nova.compute.manager [req-f36adf72-f45b-4f7f-a0d2-1067f89fb91e req-56ca3578-d362-41a1-a174-922238ae9a74 service nova] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Received event network-vif-plugged-844976e8-2627-499f-9b01-284d376c86fa {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 826.878453] env[62385]: DEBUG oslo_concurrency.lockutils [req-f36adf72-f45b-4f7f-a0d2-1067f89fb91e req-56ca3578-d362-41a1-a174-922238ae9a74 service nova] Acquiring lock "5a82fc98-e656-4476-b7dc-99466e9d6afd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.878775] env[62385]: DEBUG oslo_concurrency.lockutils [req-f36adf72-f45b-4f7f-a0d2-1067f89fb91e req-56ca3578-d362-41a1-a174-922238ae9a74 service nova] Lock "5a82fc98-e656-4476-b7dc-99466e9d6afd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.879045] env[62385]: DEBUG oslo_concurrency.lockutils [req-f36adf72-f45b-4f7f-a0d2-1067f89fb91e req-56ca3578-d362-41a1-a174-922238ae9a74 service nova] Lock "5a82fc98-e656-4476-b7dc-99466e9d6afd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.879207] env[62385]: DEBUG nova.compute.manager [req-f36adf72-f45b-4f7f-a0d2-1067f89fb91e req-56ca3578-d362-41a1-a174-922238ae9a74 service nova] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] No waiting events found dispatching network-vif-plugged-844976e8-2627-499f-9b01-284d376c86fa {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 826.879460] env[62385]: WARNING nova.compute.manager [req-f36adf72-f45b-4f7f-a0d2-1067f89fb91e req-56ca3578-d362-41a1-a174-922238ae9a74 service nova] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Received unexpected event network-vif-plugged-844976e8-2627-499f-9b01-284d376c86fa for instance with vm_state building and task_state spawning. [ 826.905129] env[62385]: DEBUG nova.network.neutron [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Successfully updated port: 844976e8-2627-499f-9b01-284d376c86fa {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 827.175131] env[62385]: DEBUG nova.scheduler.client.report [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Updated inventory for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a with generation 86 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 827.175477] env[62385]: DEBUG nova.compute.provider_tree [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Updating resource provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a generation from 86 to 87 during operation: update_inventory {{(pid=62385) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 827.175991] env[62385]: DEBUG nova.compute.provider_tree [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Updating inventory in ProviderTree for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 827.198942] env[62385]: DEBUG oslo_concurrency.lockutils [None req-526d9d0e-7639-4afe-a536-ccf937e01f2d tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "interface-0d823fcc-cbba-41f6-a47c-2f57d3e79948-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.199541] env[62385]: DEBUG oslo_concurrency.lockutils [None req-526d9d0e-7639-4afe-a536-ccf937e01f2d tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "interface-0d823fcc-cbba-41f6-a47c-2f57d3e79948-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.200362] env[62385]: DEBUG nova.objects.instance [None req-526d9d0e-7639-4afe-a536-ccf937e01f2d tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lazy-loading 'flavor' on Instance uuid 0d823fcc-cbba-41f6-a47c-2f57d3e79948 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 827.214650] env[62385]: DEBUG oslo_vmware.api [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205823, 'name': PowerOnVM_Task, 'duration_secs': 0.505065} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.215046] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 827.215368] env[62385]: INFO nova.compute.manager [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Took 6.76 seconds to spawn the instance on the hypervisor. [ 827.216077] env[62385]: DEBUG nova.compute.manager [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 827.217817] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33b97090-63da-4060-8871-330a663f8d7e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.304629] env[62385]: DEBUG oslo_concurrency.lockutils [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.414567] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "refresh_cache-5a82fc98-e656-4476-b7dc-99466e9d6afd" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.414567] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquired lock "refresh_cache-5a82fc98-e656-4476-b7dc-99466e9d6afd" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.414567] env[62385]: DEBUG nova.network.neutron [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 827.471545] env[62385]: DEBUG nova.compute.manager [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 827.472982] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af39f375-13ac-41cd-861e-c693e6884449 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.681878] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.711s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.685136] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.116s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.686747] env[62385]: INFO nova.compute.claims [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 827.742657] env[62385]: INFO nova.compute.manager [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Took 40.61 seconds to build instance. [ 827.833701] env[62385]: DEBUG nova.objects.instance [None req-526d9d0e-7639-4afe-a536-ccf937e01f2d tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lazy-loading 'pci_requests' on Instance uuid 0d823fcc-cbba-41f6-a47c-2f57d3e79948 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 827.963930] env[62385]: DEBUG nova.network.neutron [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 827.986054] env[62385]: INFO nova.compute.manager [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] instance snapshotting [ 827.986456] env[62385]: WARNING nova.compute.manager [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] trying to snapshot a non-running instance: (state: 7 expected: 1) [ 827.993816] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a813ae12-7ac9-4972-81c0-7ac792fccb13 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.023185] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1ec5e36-f24b-4388-b633-9891b8a49cec {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.178239] env[62385]: DEBUG nova.network.neutron [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Updating instance_info_cache with network_info: [{"id": "844976e8-2627-499f-9b01-284d376c86fa", "address": "fa:16:3e:89:10:4d", "network": {"id": "30c4e78b-6a64-47dc-aaaa-a210160c9f23", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1809587656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "105150c170d74e268e82ab894e9bf0d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10ff2092-e8eb-4768-ad4a-65a80560b447", "external-id": "nsx-vlan-transportzone-845", "segmentation_id": 845, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap844976e8-26", "ovs_interfaceid": "844976e8-2627-499f-9b01-284d376c86fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.199281] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d358af22-5eb5-44e3-92ce-67f5fa5e1296 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "f8a98710-9d02-4aa3-b6a5-34b7060ad62d" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 52.488s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.200174] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3e237177-68bb-4b96-9885-fed59468a6c1 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "f8a98710-9d02-4aa3-b6a5-34b7060ad62d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 32.642s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.200403] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3e237177-68bb-4b96-9885-fed59468a6c1 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "f8a98710-9d02-4aa3-b6a5-34b7060ad62d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.200613] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3e237177-68bb-4b96-9885-fed59468a6c1 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "f8a98710-9d02-4aa3-b6a5-34b7060ad62d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.200783] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3e237177-68bb-4b96-9885-fed59468a6c1 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "f8a98710-9d02-4aa3-b6a5-34b7060ad62d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.203052] env[62385]: INFO nova.compute.manager [None req-3e237177-68bb-4b96-9885-fed59468a6c1 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Terminating instance [ 828.206265] env[62385]: DEBUG nova.compute.manager [None req-3e237177-68bb-4b96-9885-fed59468a6c1 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 828.206478] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3e237177-68bb-4b96-9885-fed59468a6c1 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 828.207403] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-daea2238-6c34-481b-9f46-2e665e99cea1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.218896] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dee7df2f-cb27-4e3e-b715-cd48d8e2df9e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.254674] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a6d1e79a-32d0-4d7c-8699-22ca21721bf6 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "679150f5-ad17-45e6-9e3d-9f6e28b1aee3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 93.556s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.255622] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-3e237177-68bb-4b96-9885-fed59468a6c1 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f8a98710-9d02-4aa3-b6a5-34b7060ad62d could not be found. [ 828.255827] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3e237177-68bb-4b96-9885-fed59468a6c1 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 828.256063] env[62385]: INFO nova.compute.manager [None req-3e237177-68bb-4b96-9885-fed59468a6c1 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Took 0.05 seconds to destroy the instance on the hypervisor. [ 828.256320] env[62385]: DEBUG oslo.service.loopingcall [None req-3e237177-68bb-4b96-9885-fed59468a6c1 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 828.256570] env[62385]: DEBUG nova.compute.manager [-] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 828.256669] env[62385]: DEBUG nova.network.neutron [-] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 828.336441] env[62385]: DEBUG nova.objects.base [None req-526d9d0e-7639-4afe-a536-ccf937e01f2d tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Object Instance<0d823fcc-cbba-41f6-a47c-2f57d3e79948> lazy-loaded attributes: flavor,pci_requests {{(pid=62385) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 828.337121] env[62385]: DEBUG nova.network.neutron [None req-526d9d0e-7639-4afe-a536-ccf937e01f2d tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 828.414111] env[62385]: DEBUG nova.policy [None req-526d9d0e-7639-4afe-a536-ccf937e01f2d tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5040b20c848c4d28a198f1773a3ff21e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '012426f098ce40c3aaa00f628fe9cebb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 828.535925] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Creating Snapshot of the VM instance {{(pid=62385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 828.536281] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-28fe738b-2e1e-4a8a-b130-3b86eee78567 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.545437] env[62385]: DEBUG oslo_vmware.api [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 828.545437] env[62385]: value = "task-1205824" [ 828.545437] env[62385]: _type = "Task" [ 828.545437] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.557019] env[62385]: DEBUG oslo_vmware.api [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205824, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.683339] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Releasing lock "refresh_cache-5a82fc98-e656-4476-b7dc-99466e9d6afd" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.683339] env[62385]: DEBUG nova.compute.manager [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Instance network_info: |[{"id": "844976e8-2627-499f-9b01-284d376c86fa", "address": "fa:16:3e:89:10:4d", "network": {"id": "30c4e78b-6a64-47dc-aaaa-a210160c9f23", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1809587656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "105150c170d74e268e82ab894e9bf0d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10ff2092-e8eb-4768-ad4a-65a80560b447", "external-id": "nsx-vlan-transportzone-845", "segmentation_id": 845, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap844976e8-26", "ovs_interfaceid": "844976e8-2627-499f-9b01-284d376c86fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 828.684157] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:89:10:4d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '10ff2092-e8eb-4768-ad4a-65a80560b447', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '844976e8-2627-499f-9b01-284d376c86fa', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 828.696237] env[62385]: DEBUG oslo.service.loopingcall [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 828.696237] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 828.696397] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d43d4fa3-0f94-4429-8a1e-74537fa48172 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.722672] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 828.722672] env[62385]: value = "task-1205825" [ 828.722672] env[62385]: _type = "Task" [ 828.722672] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.733552] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205825, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.808257] env[62385]: DEBUG nova.network.neutron [None req-526d9d0e-7639-4afe-a536-ccf937e01f2d tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Successfully created port: 51428619-2fcb-4170-97f7-e6b126268421 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 829.005223] env[62385]: DEBUG nova.network.neutron [-] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.062363] env[62385]: DEBUG oslo_vmware.api [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205824, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.151850] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d9966c4-7860-4592-ab73-cf74b11b6e7d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.162281] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02f45724-c8e2-4c50-909c-6f165080786f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.199791] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff1ed9b4-2679-4965-9597-c1bc025808f3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.209602] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-751b60be-8deb-4207-a82e-070ee4b30f1c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.224623] env[62385]: DEBUG nova.compute.provider_tree [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 829.236586] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205825, 'name': CreateVM_Task, 'duration_secs': 0.416163} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.237725] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 829.238725] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.238950] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.239402] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 829.240022] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a42d2d8-4a56-4879-b509-2a07232ede0a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.248875] env[62385]: DEBUG oslo_vmware.api [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 829.248875] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c68a04-bba7-3d20-7238-98ab75584dc1" [ 829.248875] env[62385]: _type = "Task" [ 829.248875] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.259188] env[62385]: DEBUG oslo_vmware.api [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c68a04-bba7-3d20-7238-98ab75584dc1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.263439] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquiring lock "b7a286a2-9a19-4ef1-8a46-6e2a04e22413" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.263698] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "b7a286a2-9a19-4ef1-8a46-6e2a04e22413" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.508169] env[62385]: INFO nova.compute.manager [-] [instance: f8a98710-9d02-4aa3-b6a5-34b7060ad62d] Took 1.25 seconds to deallocate network for instance. [ 829.559327] env[62385]: DEBUG oslo_vmware.api [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205824, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.731057] env[62385]: DEBUG nova.scheduler.client.report [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 829.763132] env[62385]: DEBUG oslo_vmware.api [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c68a04-bba7-3d20-7238-98ab75584dc1, 'name': SearchDatastore_Task, 'duration_secs': 0.02094} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.763132] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.763132] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 829.763132] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.763596] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.763596] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 829.765031] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d8503d79-a729-472d-acd5-170b5dc26e2f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.767734] env[62385]: DEBUG nova.compute.manager [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 829.781741] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 829.781959] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 829.782782] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-63cc3285-ea24-4dd1-904e-97847ec8dce1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.789894] env[62385]: DEBUG oslo_vmware.api [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 829.789894] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]525bde01-f9ca-1c85-36bc-07a68a1f9621" [ 829.789894] env[62385]: _type = "Task" [ 829.789894] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.799678] env[62385]: DEBUG oslo_vmware.api [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]525bde01-f9ca-1c85-36bc-07a68a1f9621, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.930222] env[62385]: DEBUG nova.compute.manager [req-92bb8fc5-6021-49a4-b3a6-d6a4e1443e30 req-6914028a-4c6a-4bb9-b422-9fab10650d37 service nova] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Received event network-changed-844976e8-2627-499f-9b01-284d376c86fa {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 829.930411] env[62385]: DEBUG nova.compute.manager [req-92bb8fc5-6021-49a4-b3a6-d6a4e1443e30 req-6914028a-4c6a-4bb9-b422-9fab10650d37 service nova] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Refreshing instance network info cache due to event network-changed-844976e8-2627-499f-9b01-284d376c86fa. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 829.930646] env[62385]: DEBUG oslo_concurrency.lockutils [req-92bb8fc5-6021-49a4-b3a6-d6a4e1443e30 req-6914028a-4c6a-4bb9-b422-9fab10650d37 service nova] Acquiring lock "refresh_cache-5a82fc98-e656-4476-b7dc-99466e9d6afd" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.930821] env[62385]: DEBUG oslo_concurrency.lockutils [req-92bb8fc5-6021-49a4-b3a6-d6a4e1443e30 req-6914028a-4c6a-4bb9-b422-9fab10650d37 service nova] Acquired lock "refresh_cache-5a82fc98-e656-4476-b7dc-99466e9d6afd" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.930978] env[62385]: DEBUG nova.network.neutron [req-92bb8fc5-6021-49a4-b3a6-d6a4e1443e30 req-6914028a-4c6a-4bb9-b422-9fab10650d37 service nova] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Refreshing network info cache for port 844976e8-2627-499f-9b01-284d376c86fa {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 830.060310] env[62385]: DEBUG oslo_vmware.api [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205824, 'name': CreateSnapshot_Task, 'duration_secs': 1.257534} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.060437] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Created Snapshot of the VM instance {{(pid=62385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 830.061307] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12c7b1da-241d-47ee-a6bc-2e3d8e58345a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.238432] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.554s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.239174] env[62385]: DEBUG nova.compute.manager [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 830.242806] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 35.275s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.255022] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Acquiring lock "fbc71ada-d3b5-46f6-90a9-489c118d5126" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.255022] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Lock "fbc71ada-d3b5-46f6-90a9-489c118d5126" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.295434] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.302659] env[62385]: DEBUG oslo_vmware.api [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]525bde01-f9ca-1c85-36bc-07a68a1f9621, 'name': SearchDatastore_Task, 'duration_secs': 0.018591} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.303709] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-44a2b953-48c7-4907-877f-71348af5566a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.314580] env[62385]: DEBUG oslo_vmware.api [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 830.314580] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5256f254-e111-eb0d-917c-603830c3fceb" [ 830.314580] env[62385]: _type = "Task" [ 830.314580] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.329755] env[62385]: DEBUG oslo_vmware.api [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5256f254-e111-eb0d-917c-603830c3fceb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.535374] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3e237177-68bb-4b96-9885-fed59468a6c1 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "f8a98710-9d02-4aa3-b6a5-34b7060ad62d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.335s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.568324] env[62385]: DEBUG nova.network.neutron [None req-526d9d0e-7639-4afe-a536-ccf937e01f2d tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Successfully updated port: 51428619-2fcb-4170-97f7-e6b126268421 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 830.585958] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Creating linked-clone VM from snapshot {{(pid=62385) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 830.589877] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-6df6e18c-d1e7-46e9-98f2-d0dd04bd6dc9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.601029] env[62385]: DEBUG oslo_vmware.api [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 830.601029] env[62385]: value = "task-1205826" [ 830.601029] env[62385]: _type = "Task" [ 830.601029] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.610444] env[62385]: DEBUG oslo_vmware.api [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205826, 'name': CloneVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.653968] env[62385]: DEBUG nova.compute.manager [req-5189ad6e-4305-4f31-a634-1e22e0abc8c0 req-c7afd8d4-45d1-4ff7-ae64-f1fb180d54f3 service nova] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Received event network-vif-plugged-51428619-2fcb-4170-97f7-e6b126268421 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 830.654026] env[62385]: DEBUG oslo_concurrency.lockutils [req-5189ad6e-4305-4f31-a634-1e22e0abc8c0 req-c7afd8d4-45d1-4ff7-ae64-f1fb180d54f3 service nova] Acquiring lock "0d823fcc-cbba-41f6-a47c-2f57d3e79948-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.654836] env[62385]: DEBUG oslo_concurrency.lockutils [req-5189ad6e-4305-4f31-a634-1e22e0abc8c0 req-c7afd8d4-45d1-4ff7-ae64-f1fb180d54f3 service nova] Lock "0d823fcc-cbba-41f6-a47c-2f57d3e79948-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.655101] env[62385]: DEBUG oslo_concurrency.lockutils [req-5189ad6e-4305-4f31-a634-1e22e0abc8c0 req-c7afd8d4-45d1-4ff7-ae64-f1fb180d54f3 service nova] Lock "0d823fcc-cbba-41f6-a47c-2f57d3e79948-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.655358] env[62385]: DEBUG nova.compute.manager [req-5189ad6e-4305-4f31-a634-1e22e0abc8c0 req-c7afd8d4-45d1-4ff7-ae64-f1fb180d54f3 service nova] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] No waiting events found dispatching network-vif-plugged-51428619-2fcb-4170-97f7-e6b126268421 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 830.655471] env[62385]: WARNING nova.compute.manager [req-5189ad6e-4305-4f31-a634-1e22e0abc8c0 req-c7afd8d4-45d1-4ff7-ae64-f1fb180d54f3 service nova] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Received unexpected event network-vif-plugged-51428619-2fcb-4170-97f7-e6b126268421 for instance with vm_state active and task_state None. [ 830.671254] env[62385]: DEBUG nova.network.neutron [req-92bb8fc5-6021-49a4-b3a6-d6a4e1443e30 req-6914028a-4c6a-4bb9-b422-9fab10650d37 service nova] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Updated VIF entry in instance network info cache for port 844976e8-2627-499f-9b01-284d376c86fa. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 830.671646] env[62385]: DEBUG nova.network.neutron [req-92bb8fc5-6021-49a4-b3a6-d6a4e1443e30 req-6914028a-4c6a-4bb9-b422-9fab10650d37 service nova] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Updating instance_info_cache with network_info: [{"id": "844976e8-2627-499f-9b01-284d376c86fa", "address": "fa:16:3e:89:10:4d", "network": {"id": "30c4e78b-6a64-47dc-aaaa-a210160c9f23", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1809587656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "105150c170d74e268e82ab894e9bf0d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10ff2092-e8eb-4768-ad4a-65a80560b447", "external-id": "nsx-vlan-transportzone-845", "segmentation_id": 845, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap844976e8-26", "ovs_interfaceid": "844976e8-2627-499f-9b01-284d376c86fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.747042] env[62385]: DEBUG nova.compute.utils [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 830.757112] env[62385]: DEBUG nova.compute.manager [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Not allocating networking since 'none' was specified. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 830.757332] env[62385]: DEBUG nova.compute.manager [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 830.825521] env[62385]: DEBUG oslo_vmware.api [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5256f254-e111-eb0d-917c-603830c3fceb, 'name': SearchDatastore_Task, 'duration_secs': 0.034667} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.825926] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.826279] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 5a82fc98-e656-4476-b7dc-99466e9d6afd/5a82fc98-e656-4476-b7dc-99466e9d6afd.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 830.826562] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-542bc6a8-a789-41b5-b973-b33985a15f77 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.835385] env[62385]: DEBUG oslo_vmware.api [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 830.835385] env[62385]: value = "task-1205827" [ 830.835385] env[62385]: _type = "Task" [ 830.835385] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.844207] env[62385]: DEBUG oslo_vmware.api [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205827, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.073764] env[62385]: DEBUG oslo_concurrency.lockutils [None req-526d9d0e-7639-4afe-a536-ccf937e01f2d tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "refresh_cache-0d823fcc-cbba-41f6-a47c-2f57d3e79948" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.073921] env[62385]: DEBUG oslo_concurrency.lockutils [None req-526d9d0e-7639-4afe-a536-ccf937e01f2d tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquired lock "refresh_cache-0d823fcc-cbba-41f6-a47c-2f57d3e79948" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.074126] env[62385]: DEBUG nova.network.neutron [None req-526d9d0e-7639-4afe-a536-ccf937e01f2d tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 831.112866] env[62385]: DEBUG oslo_vmware.api [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205826, 'name': CloneVM_Task} progress is 94%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.175243] env[62385]: DEBUG oslo_concurrency.lockutils [req-92bb8fc5-6021-49a4-b3a6-d6a4e1443e30 req-6914028a-4c6a-4bb9-b422-9fab10650d37 service nova] Releasing lock "refresh_cache-5a82fc98-e656-4476-b7dc-99466e9d6afd" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.258132] env[62385]: DEBUG nova.compute.manager [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 831.288325] env[62385]: WARNING nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 1ddb4d6d-3872-49db-bb40-e21721241e89 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 831.288325] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 33387505-c576-488b-8c9c-b064fe81a7d7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 831.288325] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance f1d33401-35f5-4d79-abb4-26dc6faa784e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 831.288564] env[62385]: WARNING nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance a5c4afc6-38a6-4815-8ec4-cc01c24489bf is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 831.288564] env[62385]: WARNING nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 65bb1d14-84b2-4e75-acdc-dc674a035101 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 831.288727] env[62385]: WARNING nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 831.288831] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 0d823fcc-cbba-41f6-a47c-2f57d3e79948 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 831.288945] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance b4970e1a-2b29-44e9-b79e-cda3c32dddcd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 831.289087] env[62385]: WARNING nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance fd7a77e2-6d76-4d91-bdef-e30333247aa9 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 831.289200] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 831.289310] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 9d6f098a-0b05-43ef-96b0-9eb99ad3538c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 831.289433] env[62385]: WARNING nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 2b744879-01d2-440e-8adf-58438b922888 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 831.289560] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance e4089e71-eea1-4e16-b90c-97966333f2bc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 831.289704] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance f2be792c-a723-4a69-b255-bbe7f0e8692d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 831.289818] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 55d62bba-ff30-42bf-b1cb-7567988a9361 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 831.289927] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 831.290048] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 679150f5-ad17-45e6-9e3d-9f6e28b1aee3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 831.290159] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 5a82fc98-e656-4476-b7dc-99466e9d6afd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 831.290274] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 7f2fe830-e16d-4684-91a4-aab219468e77 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 831.292626] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.346316] env[62385]: DEBUG oslo_vmware.api [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205827, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.551188] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "fe0624a9-09a8-498b-bb3c-fda6cab92341" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.551427] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "fe0624a9-09a8-498b-bb3c-fda6cab92341" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.612299] env[62385]: DEBUG oslo_vmware.api [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205826, 'name': CloneVM_Task} progress is 94%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.613803] env[62385]: WARNING nova.network.neutron [None req-526d9d0e-7639-4afe-a536-ccf937e01f2d tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] daad7070-dc06-4580-9f4a-ac8aaa546955 already exists in list: networks containing: ['daad7070-dc06-4580-9f4a-ac8aaa546955']. ignoring it [ 831.796048] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance b73a31df-53c8-4550-bf75-5cf3b5aff86c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 831.848949] env[62385]: DEBUG oslo_vmware.api [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205827, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.632845} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.852961] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 5a82fc98-e656-4476-b7dc-99466e9d6afd/5a82fc98-e656-4476-b7dc-99466e9d6afd.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 831.853518] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 831.853948] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d42762e2-baf7-41b1-98c3-9bc177d59a1d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.864749] env[62385]: DEBUG oslo_vmware.api [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 831.864749] env[62385]: value = "task-1205828" [ 831.864749] env[62385]: _type = "Task" [ 831.864749] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.873863] env[62385]: DEBUG oslo_vmware.api [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205828, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.936742] env[62385]: DEBUG nova.network.neutron [None req-526d9d0e-7639-4afe-a536-ccf937e01f2d tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Updating instance_info_cache with network_info: [{"id": "ad929652-f520-465a-aeea-1db4e8f36fc3", "address": "fa:16:3e:b6:00:7c", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.130", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad929652-f5", "ovs_interfaceid": "ad929652-f520-465a-aeea-1db4e8f36fc3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "51428619-2fcb-4170-97f7-e6b126268421", "address": "fa:16:3e:ae:77:7b", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51428619-2f", "ovs_interfaceid": "51428619-2fcb-4170-97f7-e6b126268421", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.054127] env[62385]: DEBUG nova.compute.manager [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 832.115285] env[62385]: DEBUG oslo_vmware.api [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205826, 'name': CloneVM_Task} progress is 95%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.274557] env[62385]: DEBUG nova.compute.manager [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 832.297576] env[62385]: DEBUG nova.virt.hardware [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 832.300166] env[62385]: DEBUG nova.virt.hardware [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 832.300166] env[62385]: DEBUG nova.virt.hardware [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 832.300166] env[62385]: DEBUG nova.virt.hardware [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 832.300166] env[62385]: DEBUG nova.virt.hardware [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 832.300166] env[62385]: DEBUG nova.virt.hardware [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 832.300166] env[62385]: DEBUG nova.virt.hardware [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 832.300166] env[62385]: DEBUG nova.virt.hardware [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 832.300166] env[62385]: DEBUG nova.virt.hardware [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 832.300166] env[62385]: DEBUG nova.virt.hardware [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 832.300166] env[62385]: DEBUG nova.virt.hardware [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 832.301083] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a10b8a2-3d31-41e3-bfc8-ffd9dc67c675 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.304616] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 2a579d13-5372-4340-b7b3-cc02c1912624 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 832.312483] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13bc396d-ac7c-4674-868e-6de06cc4c958 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.330574] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Instance VIF info [] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 832.341429] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Creating folder: Project (3c86d0415efb4ea880e15a8bfa0a56bc). Parent ref: group-v261107. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 832.343338] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6492b18c-13cb-479f-bc83-c4b926760b03 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.354793] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Created folder: Project (3c86d0415efb4ea880e15a8bfa0a56bc) in parent group-v261107. [ 832.355194] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Creating folder: Instances. Parent ref: group-v261211. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 832.355766] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cbc91798-8c2a-4d2d-8a33-9839dd09b890 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.366715] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Created folder: Instances in parent group-v261211. [ 832.367049] env[62385]: DEBUG oslo.service.loopingcall [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 832.371143] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 832.371476] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1ec86be9-52d9-40be-b3c9-ef25ec154599 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.389613] env[62385]: DEBUG oslo_vmware.api [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205828, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.095679} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.391025] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 832.391349] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 832.391349] env[62385]: value = "task-1205831" [ 832.391349] env[62385]: _type = "Task" [ 832.391349] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.392195] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90c891b2-fb4f-4c41-b8c4-a12860d91dd0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.404568] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205831, 'name': CreateVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.426191] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] 5a82fc98-e656-4476-b7dc-99466e9d6afd/5a82fc98-e656-4476-b7dc-99466e9d6afd.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 832.426983] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f7586b7c-a1ce-4b10-945b-70610a5abde0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.443612] env[62385]: DEBUG oslo_concurrency.lockutils [None req-526d9d0e-7639-4afe-a536-ccf937e01f2d tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Releasing lock "refresh_cache-0d823fcc-cbba-41f6-a47c-2f57d3e79948" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.444738] env[62385]: DEBUG oslo_concurrency.lockutils [None req-526d9d0e-7639-4afe-a536-ccf937e01f2d tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "0d823fcc-cbba-41f6-a47c-2f57d3e79948" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.445152] env[62385]: DEBUG oslo_concurrency.lockutils [None req-526d9d0e-7639-4afe-a536-ccf937e01f2d tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquired lock "0d823fcc-cbba-41f6-a47c-2f57d3e79948" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.446645] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea458f8c-f2e6-48fa-ac2f-44a244e1f1af {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.464765] env[62385]: DEBUG nova.virt.hardware [None req-526d9d0e-7639-4afe-a536-ccf937e01f2d tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 832.465102] env[62385]: DEBUG nova.virt.hardware [None req-526d9d0e-7639-4afe-a536-ccf937e01f2d tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 832.465311] env[62385]: DEBUG nova.virt.hardware [None req-526d9d0e-7639-4afe-a536-ccf937e01f2d tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 832.465559] env[62385]: DEBUG nova.virt.hardware [None req-526d9d0e-7639-4afe-a536-ccf937e01f2d tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 832.465753] env[62385]: DEBUG nova.virt.hardware [None req-526d9d0e-7639-4afe-a536-ccf937e01f2d tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 832.465944] env[62385]: DEBUG nova.virt.hardware [None req-526d9d0e-7639-4afe-a536-ccf937e01f2d tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 832.466213] env[62385]: DEBUG nova.virt.hardware [None req-526d9d0e-7639-4afe-a536-ccf937e01f2d tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 832.466434] env[62385]: DEBUG nova.virt.hardware [None req-526d9d0e-7639-4afe-a536-ccf937e01f2d tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 832.466712] env[62385]: DEBUG nova.virt.hardware [None req-526d9d0e-7639-4afe-a536-ccf937e01f2d tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 832.467176] env[62385]: DEBUG nova.virt.hardware [None req-526d9d0e-7639-4afe-a536-ccf937e01f2d tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 832.467176] env[62385]: DEBUG nova.virt.hardware [None req-526d9d0e-7639-4afe-a536-ccf937e01f2d tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 832.473811] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-526d9d0e-7639-4afe-a536-ccf937e01f2d tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Reconfiguring VM to attach interface {{(pid=62385) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 832.478027] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e95e12ef-39f1-4276-93b0-f29a86fc8f5b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.490752] env[62385]: DEBUG oslo_vmware.api [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 832.490752] env[62385]: value = "task-1205832" [ 832.490752] env[62385]: _type = "Task" [ 832.490752] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.496283] env[62385]: DEBUG oslo_vmware.api [None req-526d9d0e-7639-4afe-a536-ccf937e01f2d tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 832.496283] env[62385]: value = "task-1205833" [ 832.496283] env[62385]: _type = "Task" [ 832.496283] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.504056] env[62385]: DEBUG oslo_vmware.api [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205832, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.509480] env[62385]: DEBUG oslo_vmware.api [None req-526d9d0e-7639-4afe-a536-ccf937e01f2d tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1205833, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.587308] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.613619] env[62385]: DEBUG oslo_vmware.api [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205826, 'name': CloneVM_Task} progress is 95%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.808512] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 832.814153] env[62385]: DEBUG nova.compute.manager [req-0ed878aa-76a0-4b80-8172-256bd1cfb9d1 req-6b0b93b1-55cf-4104-8b66-94017f71a652 service nova] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Received event network-changed-51428619-2fcb-4170-97f7-e6b126268421 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 832.814153] env[62385]: DEBUG nova.compute.manager [req-0ed878aa-76a0-4b80-8172-256bd1cfb9d1 req-6b0b93b1-55cf-4104-8b66-94017f71a652 service nova] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Refreshing instance network info cache due to event network-changed-51428619-2fcb-4170-97f7-e6b126268421. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 832.814404] env[62385]: DEBUG oslo_concurrency.lockutils [req-0ed878aa-76a0-4b80-8172-256bd1cfb9d1 req-6b0b93b1-55cf-4104-8b66-94017f71a652 service nova] Acquiring lock "refresh_cache-0d823fcc-cbba-41f6-a47c-2f57d3e79948" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.814685] env[62385]: DEBUG oslo_concurrency.lockutils [req-0ed878aa-76a0-4b80-8172-256bd1cfb9d1 req-6b0b93b1-55cf-4104-8b66-94017f71a652 service nova] Acquired lock "refresh_cache-0d823fcc-cbba-41f6-a47c-2f57d3e79948" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.814685] env[62385]: DEBUG nova.network.neutron [req-0ed878aa-76a0-4b80-8172-256bd1cfb9d1 req-6b0b93b1-55cf-4104-8b66-94017f71a652 service nova] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Refreshing network info cache for port 51428619-2fcb-4170-97f7-e6b126268421 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 832.905279] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205831, 'name': CreateVM_Task, 'duration_secs': 0.357905} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.905489] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 832.905968] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.906153] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.906521] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 832.906797] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a96e7779-a24d-42b2-962e-1eceb580bb15 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.911784] env[62385]: DEBUG oslo_vmware.api [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Waiting for the task: (returnval){ [ 832.911784] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52d29d1d-eeea-9b63-0fa1-7f2b381d4c99" [ 832.911784] env[62385]: _type = "Task" [ 832.911784] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.919721] env[62385]: DEBUG oslo_vmware.api [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52d29d1d-eeea-9b63-0fa1-7f2b381d4c99, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.004344] env[62385]: DEBUG oslo_vmware.api [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205832, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.013500] env[62385]: DEBUG oslo_vmware.api [None req-526d9d0e-7639-4afe-a536-ccf937e01f2d tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1205833, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.112650] env[62385]: DEBUG oslo_vmware.api [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205826, 'name': CloneVM_Task, 'duration_secs': 2.186862} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.112956] env[62385]: INFO nova.virt.vmwareapi.vmops [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Created linked-clone VM from snapshot [ 833.113771] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5373cfd-0fad-4150-a0b8-7f8b013e3afd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.121789] env[62385]: DEBUG nova.virt.vmwareapi.images [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Uploading image 20957960-bb46-48a4-9b5c-481d6c93295a {{(pid=62385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 833.146368] env[62385]: DEBUG oslo_vmware.rw_handles [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 833.146368] env[62385]: value = "vm-261210" [ 833.146368] env[62385]: _type = "VirtualMachine" [ 833.146368] env[62385]: }. {{(pid=62385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 833.146760] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-ce2ff99e-0574-4ca5-89fd-5c3e49ee0e11 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.154582] env[62385]: DEBUG oslo_vmware.rw_handles [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lease: (returnval){ [ 833.154582] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]525333df-8b4c-cb63-34bf-d8297edcec20" [ 833.154582] env[62385]: _type = "HttpNfcLease" [ 833.154582] env[62385]: } obtained for exporting VM: (result){ [ 833.154582] env[62385]: value = "vm-261210" [ 833.154582] env[62385]: _type = "VirtualMachine" [ 833.154582] env[62385]: }. {{(pid=62385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 833.154958] env[62385]: DEBUG oslo_vmware.api [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the lease: (returnval){ [ 833.154958] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]525333df-8b4c-cb63-34bf-d8297edcec20" [ 833.154958] env[62385]: _type = "HttpNfcLease" [ 833.154958] env[62385]: } to be ready. {{(pid=62385) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 833.164612] env[62385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 833.164612] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]525333df-8b4c-cb63-34bf-d8297edcec20" [ 833.164612] env[62385]: _type = "HttpNfcLease" [ 833.164612] env[62385]: } is initializing. {{(pid=62385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 833.312252] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 699a57a9-5a1a-4cd1-8449-723400d9caf4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 833.424034] env[62385]: DEBUG oslo_vmware.api [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52d29d1d-eeea-9b63-0fa1-7f2b381d4c99, 'name': SearchDatastore_Task, 'duration_secs': 0.014359} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.424280] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.424530] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 833.424770] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.424921] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.425120] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 833.425400] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-66b94221-5434-408c-94b6-8f070f36c604 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.434042] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 833.434232] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 833.437050] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d3761e6-fe37-4332-b948-d19aeb4c596a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.443596] env[62385]: DEBUG oslo_vmware.api [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Waiting for the task: (returnval){ [ 833.443596] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52176a4b-8be9-e403-d2a1-ef2dd1bc4922" [ 833.443596] env[62385]: _type = "Task" [ 833.443596] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.451558] env[62385]: DEBUG oslo_vmware.api [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52176a4b-8be9-e403-d2a1-ef2dd1bc4922, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.508847] env[62385]: DEBUG oslo_vmware.api [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205832, 'name': ReconfigVM_Task, 'duration_secs': 0.558755} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.513087] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Reconfigured VM instance instance-00000046 to attach disk [datastore2] 5a82fc98-e656-4476-b7dc-99466e9d6afd/5a82fc98-e656-4476-b7dc-99466e9d6afd.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 833.513782] env[62385]: DEBUG oslo_vmware.api [None req-526d9d0e-7639-4afe-a536-ccf937e01f2d tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1205833, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.514019] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bd5fe600-e12c-4940-85e8-3fc9a59392de {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.521640] env[62385]: DEBUG oslo_vmware.api [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 833.521640] env[62385]: value = "task-1205835" [ 833.521640] env[62385]: _type = "Task" [ 833.521640] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.530749] env[62385]: DEBUG oslo_vmware.api [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205835, 'name': Rename_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.574218] env[62385]: DEBUG nova.network.neutron [req-0ed878aa-76a0-4b80-8172-256bd1cfb9d1 req-6b0b93b1-55cf-4104-8b66-94017f71a652 service nova] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Updated VIF entry in instance network info cache for port 51428619-2fcb-4170-97f7-e6b126268421. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 833.574720] env[62385]: DEBUG nova.network.neutron [req-0ed878aa-76a0-4b80-8172-256bd1cfb9d1 req-6b0b93b1-55cf-4104-8b66-94017f71a652 service nova] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Updating instance_info_cache with network_info: [{"id": "ad929652-f520-465a-aeea-1db4e8f36fc3", "address": "fa:16:3e:b6:00:7c", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.130", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad929652-f5", "ovs_interfaceid": "ad929652-f520-465a-aeea-1db4e8f36fc3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "51428619-2fcb-4170-97f7-e6b126268421", "address": "fa:16:3e:ae:77:7b", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51428619-2f", "ovs_interfaceid": "51428619-2fcb-4170-97f7-e6b126268421", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.663543] env[62385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 833.663543] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]525333df-8b4c-cb63-34bf-d8297edcec20" [ 833.663543] env[62385]: _type = "HttpNfcLease" [ 833.663543] env[62385]: } is ready. {{(pid=62385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 833.663906] env[62385]: DEBUG oslo_vmware.rw_handles [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 833.663906] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]525333df-8b4c-cb63-34bf-d8297edcec20" [ 833.663906] env[62385]: _type = "HttpNfcLease" [ 833.663906] env[62385]: }. {{(pid=62385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 833.664765] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d112547e-bd2d-4e98-8d70-64f52f3ff2f4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.672956] env[62385]: DEBUG oslo_vmware.rw_handles [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52052f20-a802-c130-fe17-39794709622d/disk-0.vmdk from lease info. {{(pid=62385) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 833.673204] env[62385]: DEBUG oslo_vmware.rw_handles [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52052f20-a802-c130-fe17-39794709622d/disk-0.vmdk for reading. {{(pid=62385) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 833.781743] env[62385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e364bdf7-3b73-4f59-ac88-94b21b664ad9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.815812] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance b7a286a2-9a19-4ef1-8a46-6e2a04e22413 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 833.955138] env[62385]: DEBUG oslo_vmware.api [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52176a4b-8be9-e403-d2a1-ef2dd1bc4922, 'name': SearchDatastore_Task, 'duration_secs': 0.011325} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.956022] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6892ac2c-cada-4765-8624-f6a410972cf7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.962522] env[62385]: DEBUG oslo_vmware.api [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Waiting for the task: (returnval){ [ 833.962522] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e0ad5c-2505-11f8-a610-fdf222ac2953" [ 833.962522] env[62385]: _type = "Task" [ 833.962522] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.970694] env[62385]: DEBUG oslo_vmware.api [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e0ad5c-2505-11f8-a610-fdf222ac2953, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.013310] env[62385]: DEBUG oslo_vmware.api [None req-526d9d0e-7639-4afe-a536-ccf937e01f2d tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1205833, 'name': ReconfigVM_Task, 'duration_secs': 1.266676} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.013940] env[62385]: DEBUG oslo_concurrency.lockutils [None req-526d9d0e-7639-4afe-a536-ccf937e01f2d tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Releasing lock "0d823fcc-cbba-41f6-a47c-2f57d3e79948" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.014257] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-526d9d0e-7639-4afe-a536-ccf937e01f2d tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Reconfigured VM to attach interface {{(pid=62385) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 834.031878] env[62385]: DEBUG oslo_vmware.api [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205835, 'name': Rename_Task, 'duration_secs': 0.165254} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.032962] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 834.033268] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1e9452ea-0869-4a0e-97ec-e7c88efd6127 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.041022] env[62385]: DEBUG oslo_vmware.api [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 834.041022] env[62385]: value = "task-1205836" [ 834.041022] env[62385]: _type = "Task" [ 834.041022] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.049408] env[62385]: DEBUG oslo_vmware.api [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205836, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.078333] env[62385]: DEBUG oslo_concurrency.lockutils [req-0ed878aa-76a0-4b80-8172-256bd1cfb9d1 req-6b0b93b1-55cf-4104-8b66-94017f71a652 service nova] Releasing lock "refresh_cache-0d823fcc-cbba-41f6-a47c-2f57d3e79948" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.319147] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance fbc71ada-d3b5-46f6-90a9-489c118d5126 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 834.474126] env[62385]: DEBUG oslo_vmware.api [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e0ad5c-2505-11f8-a610-fdf222ac2953, 'name': SearchDatastore_Task, 'duration_secs': 0.016551} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.474444] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.474721] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 7f2fe830-e16d-4684-91a4-aab219468e77/7f2fe830-e16d-4684-91a4-aab219468e77.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 834.474995] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ddd7c299-ee68-476c-a5c3-4aca10b6c0b0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.482730] env[62385]: DEBUG oslo_vmware.api [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Waiting for the task: (returnval){ [ 834.482730] env[62385]: value = "task-1205837" [ 834.482730] env[62385]: _type = "Task" [ 834.482730] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.491530] env[62385]: DEBUG oslo_vmware.api [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': task-1205837, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.519699] env[62385]: DEBUG oslo_concurrency.lockutils [None req-526d9d0e-7639-4afe-a536-ccf937e01f2d tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "interface-0d823fcc-cbba-41f6-a47c-2f57d3e79948-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.320s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.555916] env[62385]: DEBUG oslo_vmware.api [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205836, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.566790] env[62385]: DEBUG oslo_vmware.rw_handles [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b395f0-8a10-8613-b1bd-bd588ab77f51/disk-0.vmdk. {{(pid=62385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 834.568293] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-883fd79c-2238-43d4-b0f6-846b869176ad {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.577289] env[62385]: DEBUG oslo_vmware.rw_handles [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b395f0-8a10-8613-b1bd-bd588ab77f51/disk-0.vmdk is in state: ready. {{(pid=62385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 834.577663] env[62385]: ERROR oslo_vmware.rw_handles [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b395f0-8a10-8613-b1bd-bd588ab77f51/disk-0.vmdk due to incomplete transfer. [ 834.579018] env[62385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-a0d785dd-e300-4992-b4e0-1519e852d7f1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.586953] env[62385]: DEBUG oslo_vmware.rw_handles [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b395f0-8a10-8613-b1bd-bd588ab77f51/disk-0.vmdk. {{(pid=62385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 834.587300] env[62385]: DEBUG nova.virt.vmwareapi.images [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Uploaded image 4c74362b-3503-4ff3-8801-a50effea5d68 to the Glance image server {{(pid=62385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 834.590903] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Destroying the VM {{(pid=62385) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 834.591522] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-1b5edf0b-836e-43ee-8076-2f2e0cf9d1ea {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.601647] env[62385]: DEBUG oslo_vmware.api [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Waiting for the task: (returnval){ [ 834.601647] env[62385]: value = "task-1205838" [ 834.601647] env[62385]: _type = "Task" [ 834.601647] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.614473] env[62385]: DEBUG oslo_vmware.api [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205838, 'name': Destroy_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.822249] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance fe0624a9-09a8-498b-bb3c-fda6cab92341 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 834.822749] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Total usable vcpus: 48, total allocated vcpus: 13 {{(pid=62385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 834.823033] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3008MB phys_disk=200GB used_disk=13GB total_vcpus=48 used_vcpus=13 pci_stats=[] {{(pid=62385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 834.998309] env[62385]: DEBUG oslo_vmware.api [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': task-1205837, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.056029] env[62385]: DEBUG oslo_vmware.api [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205836, 'name': PowerOnVM_Task, 'duration_secs': 0.860673} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.056595] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 835.056893] env[62385]: INFO nova.compute.manager [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Took 9.05 seconds to spawn the instance on the hypervisor. [ 835.057327] env[62385]: DEBUG nova.compute.manager [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 835.058283] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d29a9788-d152-49a5-82d5-9bb88b17b271 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.113889] env[62385]: DEBUG oslo_vmware.api [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205838, 'name': Destroy_Task} progress is 33%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.254223] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5abb0c7-01fb-433e-b601-ec3812eb9437 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.264192] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b6bd2ae-bd91-4c0d-aa27-d510075d406d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.303397] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dabcd885-99c4-4ad2-81b7-980e0802cacb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.313975] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59e59429-3d0f-42b3-b479-c7c6f4cd8a13 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.328816] env[62385]: DEBUG nova.compute.provider_tree [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 835.497669] env[62385]: DEBUG oslo_vmware.api [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': task-1205837, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.584108] env[62385]: INFO nova.compute.manager [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Took 42.07 seconds to build instance. [ 835.612077] env[62385]: DEBUG oslo_vmware.api [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205838, 'name': Destroy_Task, 'duration_secs': 0.627325} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.613060] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Destroyed the VM [ 835.613060] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Deleting Snapshot of the VM instance {{(pid=62385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 835.613185] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-e8077f07-767c-415c-b0fa-db5ee8f127f0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.620210] env[62385]: DEBUG oslo_vmware.api [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Waiting for the task: (returnval){ [ 835.620210] env[62385]: value = "task-1205839" [ 835.620210] env[62385]: _type = "Task" [ 835.620210] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.628427] env[62385]: DEBUG oslo_vmware.api [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205839, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.836091] env[62385]: DEBUG nova.scheduler.client.report [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 835.998753] env[62385]: DEBUG oslo_vmware.api [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': task-1205837, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.086922] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a7f8bfd6-5839-4c6b-a5ca-e73395c16f9e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "5a82fc98-e656-4476-b7dc-99466e9d6afd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 81.606s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.130883] env[62385]: DEBUG oslo_vmware.api [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205839, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.341935] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62385) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 836.342238] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 6.099s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.342540] env[62385]: DEBUG oslo_concurrency.lockutils [None req-650314b1-bfcb-451d-9c85-40f84a6535ea tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 41.155s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.342774] env[62385]: DEBUG oslo_concurrency.lockutils [None req-650314b1-bfcb-451d-9c85-40f84a6535ea tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.345141] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 36.787s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.345410] env[62385]: DEBUG nova.objects.instance [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62385) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 836.355167] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3e7669c8-b9b4-49bf-88ac-b13f27909a20 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "interface-0d823fcc-cbba-41f6-a47c-2f57d3e79948-51428619-2fcb-4170-97f7-e6b126268421" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.356848] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3e7669c8-b9b4-49bf-88ac-b13f27909a20 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "interface-0d823fcc-cbba-41f6-a47c-2f57d3e79948-51428619-2fcb-4170-97f7-e6b126268421" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.002s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.374330] env[62385]: INFO nova.scheduler.client.report [None req-650314b1-bfcb-451d-9c85-40f84a6535ea tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Deleted allocations for instance 65bb1d14-84b2-4e75-acdc-dc674a035101 [ 836.500079] env[62385]: DEBUG oslo_vmware.api [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': task-1205837, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.784686} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.500404] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 7f2fe830-e16d-4684-91a4-aab219468e77/7f2fe830-e16d-4684-91a4-aab219468e77.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 836.500640] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 836.500905] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5bc9f4a2-7b3c-4686-b5ca-0f445cfe38d7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.508829] env[62385]: DEBUG oslo_vmware.api [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Waiting for the task: (returnval){ [ 836.508829] env[62385]: value = "task-1205840" [ 836.508829] env[62385]: _type = "Task" [ 836.508829] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.517569] env[62385]: DEBUG oslo_vmware.api [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': task-1205840, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.631112] env[62385]: DEBUG oslo_vmware.api [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205839, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.862020] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3e7669c8-b9b4-49bf-88ac-b13f27909a20 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "0d823fcc-cbba-41f6-a47c-2f57d3e79948" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.862020] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3e7669c8-b9b4-49bf-88ac-b13f27909a20 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquired lock "0d823fcc-cbba-41f6-a47c-2f57d3e79948" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.862020] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6505eb83-9b9f-463b-8e2d-2fdf7b3f0fcb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.882454] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bd300f8-12f9-411b-8710-85c2ce65bf19 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.918414] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3e7669c8-b9b4-49bf-88ac-b13f27909a20 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Reconfiguring VM to detach interface {{(pid=62385) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 836.920016] env[62385]: DEBUG oslo_concurrency.lockutils [None req-650314b1-bfcb-451d-9c85-40f84a6535ea tempest-ServerRescueTestJSONUnderV235-1448325908 tempest-ServerRescueTestJSONUnderV235-1448325908-project-member] Lock "65bb1d14-84b2-4e75-acdc-dc674a035101" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 44.682s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.920745] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f7baa56d-3629-4120-b8f9-3517b9296b44 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.943470] env[62385]: DEBUG oslo_vmware.api [None req-3e7669c8-b9b4-49bf-88ac-b13f27909a20 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 836.943470] env[62385]: value = "task-1205841" [ 836.943470] env[62385]: _type = "Task" [ 836.943470] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.954071] env[62385]: DEBUG oslo_vmware.api [None req-3e7669c8-b9b4-49bf-88ac-b13f27909a20 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1205841, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.021192] env[62385]: DEBUG oslo_vmware.api [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': task-1205840, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.125867} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.021531] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 837.022343] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f02653d6-492c-40e0-813c-f031a8f33d21 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.042294] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] 7f2fe830-e16d-4684-91a4-aab219468e77/7f2fe830-e16d-4684-91a4-aab219468e77.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 837.042610] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8ac337d1-7e05-429a-a616-23270cbf5c6b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.062805] env[62385]: DEBUG oslo_vmware.api [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Waiting for the task: (returnval){ [ 837.062805] env[62385]: value = "task-1205842" [ 837.062805] env[62385]: _type = "Task" [ 837.062805] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.073455] env[62385]: INFO nova.compute.manager [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Rebuilding instance [ 837.074969] env[62385]: DEBUG oslo_vmware.api [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': task-1205842, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.118844] env[62385]: DEBUG nova.compute.manager [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 837.119473] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da561559-29bb-4796-afc3-5ac2a8f84ca7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.132718] env[62385]: DEBUG oslo_vmware.api [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205839, 'name': RemoveSnapshot_Task} progress is 81%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.358027] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0f7dc0eb-96ea-4d19-a8ea-4f5072cb61e0 tempest-ServersAdmin275Test-1074904668 tempest-ServersAdmin275Test-1074904668-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.358027] env[62385]: DEBUG oslo_concurrency.lockutils [None req-aedae9b2-11c8-443b-9154-3314c6ae37ea tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 37.199s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.358216] env[62385]: DEBUG oslo_concurrency.lockutils [None req-aedae9b2-11c8-443b-9154-3314c6ae37ea tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.361474] env[62385]: DEBUG oslo_concurrency.lockutils [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 35.782s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.367020] env[62385]: INFO nova.compute.claims [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 837.387593] env[62385]: INFO nova.scheduler.client.report [None req-aedae9b2-11c8-443b-9154-3314c6ae37ea tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Deleted allocations for instance fd7a77e2-6d76-4d91-bdef-e30333247aa9 [ 837.461329] env[62385]: DEBUG oslo_vmware.api [None req-3e7669c8-b9b4-49bf-88ac-b13f27909a20 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1205841, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.575296] env[62385]: DEBUG oslo_vmware.api [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': task-1205842, 'name': ReconfigVM_Task, 'duration_secs': 0.441912} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.576213] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Reconfigured VM instance instance-00000047 to attach disk [datastore2] 7f2fe830-e16d-4684-91a4-aab219468e77/7f2fe830-e16d-4684-91a4-aab219468e77.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 837.577181] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e4564f86-430f-4e3c-ab50-daeb7b55e63a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.585204] env[62385]: DEBUG oslo_vmware.api [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Waiting for the task: (returnval){ [ 837.585204] env[62385]: value = "task-1205843" [ 837.585204] env[62385]: _type = "Task" [ 837.585204] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.597068] env[62385]: DEBUG oslo_vmware.api [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': task-1205843, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.634997] env[62385]: DEBUG oslo_vmware.api [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205839, 'name': RemoveSnapshot_Task, 'duration_secs': 1.560766} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.635617] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Deleted Snapshot of the VM instance {{(pid=62385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 837.637290] env[62385]: INFO nova.compute.manager [None req-d51edfd7-5bb3-47af-bb62-f769f2deaa71 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Took 16.17 seconds to snapshot the instance on the hypervisor. [ 837.642135] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 837.645143] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-033c3621-8361-4355-8d42-21a63a59b117 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.654316] env[62385]: DEBUG oslo_vmware.api [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 837.654316] env[62385]: value = "task-1205844" [ 837.654316] env[62385]: _type = "Task" [ 837.654316] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.663962] env[62385]: DEBUG oslo_vmware.api [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205844, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.897630] env[62385]: DEBUG oslo_concurrency.lockutils [None req-aedae9b2-11c8-443b-9154-3314c6ae37ea tempest-VolumesAdminNegativeTest-1179907164 tempest-VolumesAdminNegativeTest-1179907164-project-member] Lock "fd7a77e2-6d76-4d91-bdef-e30333247aa9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 40.688s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.956475] env[62385]: DEBUG oslo_vmware.api [None req-3e7669c8-b9b4-49bf-88ac-b13f27909a20 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1205841, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.102564] env[62385]: DEBUG oslo_vmware.api [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': task-1205843, 'name': Rename_Task, 'duration_secs': 0.174295} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.103243] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 838.103776] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3c6249e8-732d-4e12-b0ef-e8ce597eb447 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.115145] env[62385]: DEBUG oslo_vmware.api [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Waiting for the task: (returnval){ [ 838.115145] env[62385]: value = "task-1205845" [ 838.115145] env[62385]: _type = "Task" [ 838.115145] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.124583] env[62385]: DEBUG oslo_vmware.api [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': task-1205845, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.163902] env[62385]: DEBUG oslo_vmware.api [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205844, 'name': PowerOffVM_Task, 'duration_secs': 0.205268} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.164272] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 838.164505] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 838.165322] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-399aeab3-ef10-418c-95fa-7ccbbe22e9e6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.175184] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 838.175519] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-35493bcd-1676-4178-8890-dde23225c403 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.294863] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 838.294863] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Deleting contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 838.295038] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Deleting the datastore file [datastore2] 5a82fc98-e656-4476-b7dc-99466e9d6afd {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 838.295599] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3bc08b88-aa44-4e17-b02c-9b47974b9782 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.303097] env[62385]: DEBUG oslo_vmware.api [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 838.303097] env[62385]: value = "task-1205847" [ 838.303097] env[62385]: _type = "Task" [ 838.303097] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.312444] env[62385]: DEBUG oslo_vmware.api [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205847, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.459294] env[62385]: DEBUG oslo_vmware.api [None req-3e7669c8-b9b4-49bf-88ac-b13f27909a20 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1205841, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.627822] env[62385]: DEBUG oslo_vmware.api [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': task-1205845, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.770726] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76e54a83-702c-4807-a6be-7bd4ca5f2ddd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.779607] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a7ab061-9f73-47d1-981b-c33d745e4291 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.827699] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a782580-5b4b-4ec8-b15c-607b1323b6c9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.837268] env[62385]: DEBUG oslo_vmware.api [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205847, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.396185} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.840127] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 838.840716] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Deleted contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 838.840716] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 838.845888] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da312c37-035d-45ef-9b91-6fe34f8859f4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.863174] env[62385]: DEBUG nova.compute.provider_tree [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 838.960867] env[62385]: DEBUG oslo_vmware.api [None req-3e7669c8-b9b4-49bf-88ac-b13f27909a20 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1205841, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.125893] env[62385]: DEBUG oslo_vmware.api [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': task-1205845, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.369323] env[62385]: DEBUG nova.scheduler.client.report [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 839.461080] env[62385]: DEBUG oslo_vmware.api [None req-3e7669c8-b9b4-49bf-88ac-b13f27909a20 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1205841, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.624753] env[62385]: DEBUG oslo_vmware.api [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': task-1205845, 'name': PowerOnVM_Task, 'duration_secs': 1.177685} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.625082] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 839.625311] env[62385]: INFO nova.compute.manager [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Took 7.35 seconds to spawn the instance on the hypervisor. [ 839.625638] env[62385]: DEBUG nova.compute.manager [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 839.626492] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e17d8959-2506-4231-89dd-1cf1715cab17 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.879359] env[62385]: DEBUG oslo_concurrency.lockutils [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.518s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.879886] env[62385]: DEBUG nova.compute.manager [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 839.884026] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2fbab2d2-44d4-4ed4-9df6-51fefeaa7131 tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 35.680s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.884026] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2fbab2d2-44d4-4ed4-9df6-51fefeaa7131 tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.886158] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5d6a954b-460d-4b98-b7c6-e84c4438007a tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.753s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.886398] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5d6a954b-460d-4b98-b7c6-e84c4438007a tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.888565] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.409s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.890121] env[62385]: INFO nova.compute.claims [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 839.903124] env[62385]: DEBUG nova.virt.hardware [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 839.903124] env[62385]: DEBUG nova.virt.hardware [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 839.903124] env[62385]: DEBUG nova.virt.hardware [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 839.905050] env[62385]: DEBUG nova.virt.hardware [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 839.905050] env[62385]: DEBUG nova.virt.hardware [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 839.905050] env[62385]: DEBUG nova.virt.hardware [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 839.905050] env[62385]: DEBUG nova.virt.hardware [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 839.905050] env[62385]: DEBUG nova.virt.hardware [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 839.905225] env[62385]: DEBUG nova.virt.hardware [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 839.907433] env[62385]: DEBUG nova.virt.hardware [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 839.907433] env[62385]: DEBUG nova.virt.hardware [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 839.907433] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f4a00b8-f7ef-4a79-afbd-565baacb0a6f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.918550] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64b826f2-afeb-4f81-bb22-36fc23c1b980 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.930276] env[62385]: INFO nova.scheduler.client.report [None req-5d6a954b-460d-4b98-b7c6-e84c4438007a tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Deleted allocations for instance 1ddb4d6d-3872-49db-bb40-e21721241e89 [ 839.934042] env[62385]: INFO nova.scheduler.client.report [None req-2fbab2d2-44d4-4ed4-9df6-51fefeaa7131 tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Deleted allocations for instance 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7 [ 839.947492] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:89:10:4d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '10ff2092-e8eb-4768-ad4a-65a80560b447', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '844976e8-2627-499f-9b01-284d376c86fa', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 839.955643] env[62385]: DEBUG oslo.service.loopingcall [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 839.956083] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 839.959324] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d26142e3-f5e5-4f69-9490-d0ed7693592a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.982286] env[62385]: DEBUG oslo_vmware.api [None req-3e7669c8-b9b4-49bf-88ac-b13f27909a20 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1205841, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.983608] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 839.983608] env[62385]: value = "task-1205848" [ 839.983608] env[62385]: _type = "Task" [ 839.983608] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.991408] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205848, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.097555] env[62385]: DEBUG nova.compute.manager [None req-a6e09c15-369b-4d28-968e-f207336024dc tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 840.098418] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f6f3235-5f88-43ac-9199-dd2dee24e0fa {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.105377] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2aecba33-8842-48e6-bc38-81bd5b2a0bec tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Acquiring lock "d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.105576] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2aecba33-8842-48e6-bc38-81bd5b2a0bec tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.150243] env[62385]: INFO nova.compute.manager [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Took 45.60 seconds to build instance. [ 840.395274] env[62385]: DEBUG nova.compute.utils [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 840.398507] env[62385]: DEBUG nova.compute.manager [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 840.398716] env[62385]: DEBUG nova.network.neutron [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 840.440212] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5d6a954b-460d-4b98-b7c6-e84c4438007a tempest-ListServerFiltersTestJSON-1035505432 tempest-ListServerFiltersTestJSON-1035505432-project-member] Lock "1ddb4d6d-3872-49db-bb40-e21721241e89" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.300s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.449763] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2fbab2d2-44d4-4ed4-9df6-51fefeaa7131 tempest-ServersAdmin275Test-325190523 tempest-ServersAdmin275Test-325190523-project-member] Lock "8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 40.474s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.455644] env[62385]: DEBUG nova.policy [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5e83312d226945c99d05fcc7f74c0978', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bf8e3c942d3445919cfbe988cca84e90', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 840.470470] env[62385]: DEBUG oslo_vmware.api [None req-3e7669c8-b9b4-49bf-88ac-b13f27909a20 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1205841, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.497112] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205848, 'name': CreateVM_Task, 'duration_secs': 0.496693} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.497112] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 840.497873] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.498106] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.498492] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 840.499217] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-80a9119a-127f-45a4-9a73-2a1e831adc5f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.505144] env[62385]: DEBUG oslo_vmware.api [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 840.505144] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]528894eb-44a3-0595-898f-724553b476b2" [ 840.505144] env[62385]: _type = "Task" [ 840.505144] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.514033] env[62385]: DEBUG oslo_vmware.api [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]528894eb-44a3-0595-898f-724553b476b2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.537569] env[62385]: INFO nova.compute.manager [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Rebuilding instance [ 840.580781] env[62385]: DEBUG nova.compute.manager [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 840.581818] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6059d989-7efc-4fe7-ad26-58ebdafca4d8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.612637] env[62385]: DEBUG nova.compute.utils [None req-2aecba33-8842-48e6-bc38-81bd5b2a0bec tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 840.615219] env[62385]: INFO nova.compute.manager [None req-a6e09c15-369b-4d28-968e-f207336024dc tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] instance snapshotting [ 840.617856] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-801e537d-c7ca-4ace-8257-8ffffe6033b8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.641711] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f94d7dfe-142f-4067-b39b-e8dc89965e01 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.655250] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7dd5b13b-19b9-4f40-bb6c-a1453e376652 tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Lock "7f2fe830-e16d-4684-91a4-aab219468e77" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.841s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.902715] env[62385]: DEBUG nova.compute.manager [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 840.912473] env[62385]: DEBUG nova.network.neutron [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Successfully created port: 508791ff-1aa3-43d6-9dd9-1aac3351b688 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 840.973437] env[62385]: DEBUG oslo_vmware.api [None req-3e7669c8-b9b4-49bf-88ac-b13f27909a20 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1205841, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.023092] env[62385]: DEBUG oslo_vmware.api [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]528894eb-44a3-0595-898f-724553b476b2, 'name': SearchDatastore_Task, 'duration_secs': 0.021719} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.027195] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.027519] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 841.027954] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.028254] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.029599] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 841.029599] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ebd44528-663d-4308-b280-af711e2bdee1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.044925] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 841.045323] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 841.046746] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f27d58a8-17e6-4f0b-939d-1ba23ae5bc72 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.055376] env[62385]: DEBUG oslo_vmware.api [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 841.055376] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c00481-5847-9c3a-1c8c-75f32fb0332e" [ 841.055376] env[62385]: _type = "Task" [ 841.055376] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.064206] env[62385]: DEBUG oslo_vmware.api [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c00481-5847-9c3a-1c8c-75f32fb0332e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.092883] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 841.093240] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dfbbcbda-2f28-44e4-86a9-c56698f59383 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.101079] env[62385]: DEBUG oslo_vmware.api [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Waiting for the task: (returnval){ [ 841.101079] env[62385]: value = "task-1205849" [ 841.101079] env[62385]: _type = "Task" [ 841.101079] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.119652] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2aecba33-8842-48e6-bc38-81bd5b2a0bec tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.014s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.120442] env[62385]: DEBUG oslo_vmware.api [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': task-1205849, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.154076] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a6e09c15-369b-4d28-968e-f207336024dc tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Creating Snapshot of the VM instance {{(pid=62385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 841.154603] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-a670997f-02c1-445c-8522-856ba0df67e4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.165496] env[62385]: DEBUG oslo_vmware.api [None req-a6e09c15-369b-4d28-968e-f207336024dc tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Waiting for the task: (returnval){ [ 841.165496] env[62385]: value = "task-1205850" [ 841.165496] env[62385]: _type = "Task" [ 841.165496] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.177637] env[62385]: DEBUG oslo_vmware.api [None req-a6e09c15-369b-4d28-968e-f207336024dc tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205850, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.336725] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3521a148-86e5-4265-b742-8ec378aba02f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.347731] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-436b82be-4048-4204-b6c8-eeb691984ca2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.393988] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a419b438-4f88-4516-9ba5-1564e690249f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.401794] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88f04b49-93d1-43be-962d-50909e184cec {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.417846] env[62385]: DEBUG nova.compute.provider_tree [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 841.475044] env[62385]: DEBUG oslo_vmware.api [None req-3e7669c8-b9b4-49bf-88ac-b13f27909a20 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1205841, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.570881] env[62385]: DEBUG oslo_vmware.api [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c00481-5847-9c3a-1c8c-75f32fb0332e, 'name': SearchDatastore_Task, 'duration_secs': 0.008673} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.571822] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-df8b25ba-2d1b-446e-8395-00a9a6ca258e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.578042] env[62385]: DEBUG oslo_vmware.api [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 841.578042] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]526df0ea-00de-fa93-26b2-92062a3d137b" [ 841.578042] env[62385]: _type = "Task" [ 841.578042] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.588950] env[62385]: DEBUG oslo_vmware.api [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]526df0ea-00de-fa93-26b2-92062a3d137b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.612303] env[62385]: DEBUG oslo_vmware.api [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': task-1205849, 'name': PowerOffVM_Task, 'duration_secs': 0.144422} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.612303] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 841.612494] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 841.613720] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c250662-5a74-42c6-be2b-c0627f87c819 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.620328] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 841.620708] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0de05317-8dda-4bf1-a79a-d98e54890b94 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.648960] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 841.649251] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Deleting contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 841.649624] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Deleting the datastore file [datastore2] 7f2fe830-e16d-4684-91a4-aab219468e77 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 841.649805] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-56de27be-dc44-49b3-ba5f-23375cbf568c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.658162] env[62385]: DEBUG oslo_vmware.api [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Waiting for the task: (returnval){ [ 841.658162] env[62385]: value = "task-1205852" [ 841.658162] env[62385]: _type = "Task" [ 841.658162] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.669254] env[62385]: DEBUG oslo_vmware.api [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': task-1205852, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.681745] env[62385]: DEBUG oslo_vmware.api [None req-a6e09c15-369b-4d28-968e-f207336024dc tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205850, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.924523] env[62385]: DEBUG nova.scheduler.client.report [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 841.929811] env[62385]: DEBUG nova.compute.manager [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 841.966100] env[62385]: DEBUG nova.virt.hardware [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 841.966559] env[62385]: DEBUG nova.virt.hardware [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 841.966921] env[62385]: DEBUG nova.virt.hardware [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 841.967253] env[62385]: DEBUG nova.virt.hardware [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 841.967550] env[62385]: DEBUG nova.virt.hardware [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 841.967931] env[62385]: DEBUG nova.virt.hardware [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 841.968289] env[62385]: DEBUG nova.virt.hardware [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 841.970017] env[62385]: DEBUG nova.virt.hardware [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 841.970017] env[62385]: DEBUG nova.virt.hardware [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 841.970017] env[62385]: DEBUG nova.virt.hardware [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 841.970017] env[62385]: DEBUG nova.virt.hardware [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 841.970963] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88c3bf2e-2254-4486-acc2-7908f5946557 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.982591] env[62385]: DEBUG oslo_vmware.api [None req-3e7669c8-b9b4-49bf-88ac-b13f27909a20 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1205841, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.986938] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41852db7-a9e5-41fa-a26e-8969e7854d65 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.092787] env[62385]: DEBUG oslo_vmware.api [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]526df0ea-00de-fa93-26b2-92062a3d137b, 'name': SearchDatastore_Task, 'duration_secs': 0.010949} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.093697] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.094662] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 5a82fc98-e656-4476-b7dc-99466e9d6afd/5a82fc98-e656-4476-b7dc-99466e9d6afd.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 842.094662] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-90915594-a192-4480-81d7-2b0f34e3a3ca {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.105148] env[62385]: DEBUG oslo_vmware.api [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 842.105148] env[62385]: value = "task-1205853" [ 842.105148] env[62385]: _type = "Task" [ 842.105148] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.113639] env[62385]: DEBUG oslo_vmware.api [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205853, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.170404] env[62385]: DEBUG oslo_vmware.api [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': task-1205852, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.094783} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.174141] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 842.174295] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Deleted contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 842.174736] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 842.183817] env[62385]: DEBUG oslo_vmware.api [None req-a6e09c15-369b-4d28-968e-f207336024dc tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205850, 'name': CreateSnapshot_Task, 'duration_secs': 0.6626} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.184092] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a6e09c15-369b-4d28-968e-f207336024dc tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Created Snapshot of the VM instance {{(pid=62385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 842.185127] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdedad44-4d65-4017-9901-6a796659769e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.210828] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2aecba33-8842-48e6-bc38-81bd5b2a0bec tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Acquiring lock "d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.210828] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2aecba33-8842-48e6-bc38-81bd5b2a0bec tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.210828] env[62385]: INFO nova.compute.manager [None req-2aecba33-8842-48e6-bc38-81bd5b2a0bec tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Attaching volume ddb60eb1-ec31-4439-b8c9-0c53408d8c97 to /dev/sdb [ 842.245989] env[62385]: DEBUG oslo_vmware.rw_handles [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52052f20-a802-c130-fe17-39794709622d/disk-0.vmdk. {{(pid=62385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 842.247171] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da7b1796-a43e-45d3-bae1-d81b1723cae6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.254210] env[62385]: DEBUG oslo_vmware.rw_handles [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52052f20-a802-c130-fe17-39794709622d/disk-0.vmdk is in state: ready. {{(pid=62385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 842.254210] env[62385]: ERROR oslo_vmware.rw_handles [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52052f20-a802-c130-fe17-39794709622d/disk-0.vmdk due to incomplete transfer. [ 842.254210] env[62385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-9689892f-8d6f-49d5-bc37-296a92a9ef0a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.261279] env[62385]: DEBUG oslo_vmware.rw_handles [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52052f20-a802-c130-fe17-39794709622d/disk-0.vmdk. {{(pid=62385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 842.261574] env[62385]: DEBUG nova.virt.vmwareapi.images [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Uploaded image 20957960-bb46-48a4-9b5c-481d6c93295a to the Glance image server {{(pid=62385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 842.263387] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Destroying the VM {{(pid=62385) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 842.263625] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-8f6ac342-4d46-4756-b366-ea57de362a9d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.266805] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-557fd806-ce2c-4c7e-bf61-bc06a0ed58b5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.274917] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-747b380d-8d1e-442a-ae16-73ba3371fa72 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.277667] env[62385]: DEBUG oslo_vmware.api [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 842.277667] env[62385]: value = "task-1205854" [ 842.277667] env[62385]: _type = "Task" [ 842.277667] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.288077] env[62385]: DEBUG oslo_vmware.api [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205854, 'name': Destroy_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.291949] env[62385]: DEBUG nova.virt.block_device [None req-2aecba33-8842-48e6-bc38-81bd5b2a0bec tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Updating existing volume attachment record: 170e0d93-c6c3-4c9e-b8a6-ecb6342f675b {{(pid=62385) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 842.435703] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.546s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.435826] env[62385]: DEBUG nova.compute.manager [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 842.439662] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4e6ebd10-fb45-4b48-ad7e-266e6c10e531 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.055s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.440284] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4e6ebd10-fb45-4b48-ad7e-266e6c10e531 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.446214] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.935s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.447485] env[62385]: INFO nova.compute.claims [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 842.473887] env[62385]: DEBUG oslo_vmware.api [None req-3e7669c8-b9b4-49bf-88ac-b13f27909a20 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1205841, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.489429] env[62385]: INFO nova.scheduler.client.report [None req-4e6ebd10-fb45-4b48-ad7e-266e6c10e531 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Deleted allocations for instance a5c4afc6-38a6-4815-8ec4-cc01c24489bf [ 842.620473] env[62385]: DEBUG oslo_vmware.api [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205853, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.648466] env[62385]: DEBUG nova.compute.manager [req-1da4e1c5-e1a9-4117-a32c-0642059ebda0 req-eecdd38f-b5a6-468f-a965-2ac382790059 service nova] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Received event network-vif-plugged-508791ff-1aa3-43d6-9dd9-1aac3351b688 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 842.648773] env[62385]: DEBUG oslo_concurrency.lockutils [req-1da4e1c5-e1a9-4117-a32c-0642059ebda0 req-eecdd38f-b5a6-468f-a965-2ac382790059 service nova] Acquiring lock "b73a31df-53c8-4550-bf75-5cf3b5aff86c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.649301] env[62385]: DEBUG oslo_concurrency.lockutils [req-1da4e1c5-e1a9-4117-a32c-0642059ebda0 req-eecdd38f-b5a6-468f-a965-2ac382790059 service nova] Lock "b73a31df-53c8-4550-bf75-5cf3b5aff86c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.651131] env[62385]: DEBUG oslo_concurrency.lockutils [req-1da4e1c5-e1a9-4117-a32c-0642059ebda0 req-eecdd38f-b5a6-468f-a965-2ac382790059 service nova] Lock "b73a31df-53c8-4550-bf75-5cf3b5aff86c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.651131] env[62385]: DEBUG nova.compute.manager [req-1da4e1c5-e1a9-4117-a32c-0642059ebda0 req-eecdd38f-b5a6-468f-a965-2ac382790059 service nova] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] No waiting events found dispatching network-vif-plugged-508791ff-1aa3-43d6-9dd9-1aac3351b688 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 842.651131] env[62385]: WARNING nova.compute.manager [req-1da4e1c5-e1a9-4117-a32c-0642059ebda0 req-eecdd38f-b5a6-468f-a965-2ac382790059 service nova] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Received unexpected event network-vif-plugged-508791ff-1aa3-43d6-9dd9-1aac3351b688 for instance with vm_state building and task_state spawning. [ 842.707309] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a6e09c15-369b-4d28-968e-f207336024dc tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Creating linked-clone VM from snapshot {{(pid=62385) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 842.707746] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-00599f06-1249-4c81-9806-90cbc8bb0982 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.721255] env[62385]: DEBUG oslo_vmware.api [None req-a6e09c15-369b-4d28-968e-f207336024dc tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Waiting for the task: (returnval){ [ 842.721255] env[62385]: value = "task-1205856" [ 842.721255] env[62385]: _type = "Task" [ 842.721255] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.732892] env[62385]: DEBUG oslo_vmware.api [None req-a6e09c15-369b-4d28-968e-f207336024dc tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205856, 'name': CloneVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.792310] env[62385]: DEBUG oslo_vmware.api [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205854, 'name': Destroy_Task, 'duration_secs': 0.337392} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.792434] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Destroyed the VM [ 842.792658] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Deleting Snapshot of the VM instance {{(pid=62385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 842.792917] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-cbda7b46-d764-44c7-82e5-25c559759200 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.802668] env[62385]: DEBUG oslo_vmware.api [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 842.802668] env[62385]: value = "task-1205859" [ 842.802668] env[62385]: _type = "Task" [ 842.802668] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.811629] env[62385]: DEBUG oslo_vmware.api [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205859, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.876491] env[62385]: DEBUG nova.network.neutron [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Successfully updated port: 508791ff-1aa3-43d6-9dd9-1aac3351b688 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 842.958021] env[62385]: DEBUG nova.compute.utils [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 842.965329] env[62385]: DEBUG nova.compute.manager [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 842.965530] env[62385]: DEBUG nova.network.neutron [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 842.982034] env[62385]: DEBUG oslo_vmware.api [None req-3e7669c8-b9b4-49bf-88ac-b13f27909a20 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1205841, 'name': ReconfigVM_Task, 'duration_secs': 5.873914} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.982034] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3e7669c8-b9b4-49bf-88ac-b13f27909a20 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Releasing lock "0d823fcc-cbba-41f6-a47c-2f57d3e79948" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.982034] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3e7669c8-b9b4-49bf-88ac-b13f27909a20 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Reconfigured VM to detach interface {{(pid=62385) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 842.998946] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4e6ebd10-fb45-4b48-ad7e-266e6c10e531 tempest-ImagesOneServerTestJSON-1825894756 tempest-ImagesOneServerTestJSON-1825894756-project-member] Lock "a5c4afc6-38a6-4815-8ec4-cc01c24489bf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.556s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.039019] env[62385]: DEBUG nova.policy [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ed4236fea9984babaf0b48de23a3570d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '212a132650584a4d866dc95004faa391', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 843.120402] env[62385]: DEBUG oslo_vmware.api [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205853, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.753397} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.120402] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 5a82fc98-e656-4476-b7dc-99466e9d6afd/5a82fc98-e656-4476-b7dc-99466e9d6afd.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 843.120402] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 843.120402] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d42d624c-24fb-4186-b075-628776c5fc7f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.128012] env[62385]: DEBUG oslo_vmware.api [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 843.128012] env[62385]: value = "task-1205860" [ 843.128012] env[62385]: _type = "Task" [ 843.128012] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.142190] env[62385]: DEBUG oslo_vmware.api [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205860, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.224487] env[62385]: DEBUG nova.virt.hardware [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 843.224763] env[62385]: DEBUG nova.virt.hardware [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 843.228289] env[62385]: DEBUG nova.virt.hardware [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 843.228675] env[62385]: DEBUG nova.virt.hardware [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 843.228867] env[62385]: DEBUG nova.virt.hardware [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 843.229134] env[62385]: DEBUG nova.virt.hardware [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 843.229319] env[62385]: DEBUG nova.virt.hardware [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 843.229478] env[62385]: DEBUG nova.virt.hardware [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 843.229676] env[62385]: DEBUG nova.virt.hardware [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 843.229880] env[62385]: DEBUG nova.virt.hardware [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 843.230139] env[62385]: DEBUG nova.virt.hardware [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 843.231270] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c1aa1cb-3d45-42cc-927f-5ae20d8e05ae {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.250970] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-713262a7-c845-4c20-96b3-b15b73ca09ad {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.255282] env[62385]: DEBUG oslo_vmware.api [None req-a6e09c15-369b-4d28-968e-f207336024dc tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205856, 'name': CloneVM_Task} progress is 94%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.267118] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Instance VIF info [] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 843.274430] env[62385]: DEBUG oslo.service.loopingcall [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 843.275012] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 843.275012] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-71b11cdd-5dfe-43b0-9a04-eeae259c398f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.295557] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 843.295557] env[62385]: value = "task-1205861" [ 843.295557] env[62385]: _type = "Task" [ 843.295557] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.311141] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205861, 'name': CreateVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.316554] env[62385]: DEBUG oslo_vmware.api [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205859, 'name': RemoveSnapshot_Task, 'duration_secs': 0.505264} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.316845] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Deleted Snapshot of the VM instance {{(pid=62385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 843.317096] env[62385]: INFO nova.compute.manager [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Took 15.33 seconds to snapshot the instance on the hypervisor. [ 843.382178] env[62385]: DEBUG oslo_concurrency.lockutils [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "refresh_cache-b73a31df-53c8-4550-bf75-5cf3b5aff86c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.384268] env[62385]: DEBUG oslo_concurrency.lockutils [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquired lock "refresh_cache-b73a31df-53c8-4550-bf75-5cf3b5aff86c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.384268] env[62385]: DEBUG nova.network.neutron [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 843.459823] env[62385]: DEBUG nova.network.neutron [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Successfully created port: 0be859b7-c0bc-435a-908c-49582f2d31da {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 843.466023] env[62385]: DEBUG nova.compute.manager [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 843.551246] env[62385]: DEBUG oslo_concurrency.lockutils [None req-71e9a3c9-f8fb-42f2-a138-6fdd3a07054e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquiring lock "55d62bba-ff30-42bf-b1cb-7567988a9361" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.551538] env[62385]: DEBUG oslo_concurrency.lockutils [None req-71e9a3c9-f8fb-42f2-a138-6fdd3a07054e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "55d62bba-ff30-42bf-b1cb-7567988a9361" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.551755] env[62385]: DEBUG oslo_concurrency.lockutils [None req-71e9a3c9-f8fb-42f2-a138-6fdd3a07054e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquiring lock "55d62bba-ff30-42bf-b1cb-7567988a9361-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.551943] env[62385]: DEBUG oslo_concurrency.lockutils [None req-71e9a3c9-f8fb-42f2-a138-6fdd3a07054e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "55d62bba-ff30-42bf-b1cb-7567988a9361-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.552145] env[62385]: DEBUG oslo_concurrency.lockutils [None req-71e9a3c9-f8fb-42f2-a138-6fdd3a07054e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "55d62bba-ff30-42bf-b1cb-7567988a9361-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.558508] env[62385]: DEBUG nova.compute.manager [req-8b7342bb-7f6d-49c7-8296-f9cb4d3a5404 req-0dc29ef5-27d7-4976-9bfa-b7294792af9d service nova] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Received event network-vif-deleted-51428619-2fcb-4170-97f7-e6b126268421 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 843.558508] env[62385]: INFO nova.compute.manager [req-8b7342bb-7f6d-49c7-8296-f9cb4d3a5404 req-0dc29ef5-27d7-4976-9bfa-b7294792af9d service nova] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Neutron deleted interface 51428619-2fcb-4170-97f7-e6b126268421; detaching it from the instance and deleting it from the info cache [ 843.558508] env[62385]: DEBUG nova.network.neutron [req-8b7342bb-7f6d-49c7-8296-f9cb4d3a5404 req-0dc29ef5-27d7-4976-9bfa-b7294792af9d service nova] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Updating instance_info_cache with network_info: [{"id": "ad929652-f520-465a-aeea-1db4e8f36fc3", "address": "fa:16:3e:b6:00:7c", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.130", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad929652-f5", "ovs_interfaceid": "ad929652-f520-465a-aeea-1db4e8f36fc3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.562028] env[62385]: INFO nova.compute.manager [None req-71e9a3c9-f8fb-42f2-a138-6fdd3a07054e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Terminating instance [ 843.565073] env[62385]: DEBUG nova.compute.manager [None req-71e9a3c9-f8fb-42f2-a138-6fdd3a07054e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 843.565377] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-71e9a3c9-f8fb-42f2-a138-6fdd3a07054e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 843.566312] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bc1d085-88e3-42b0-897c-c5504f7251e8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.576932] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-71e9a3c9-f8fb-42f2-a138-6fdd3a07054e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 843.581217] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a211649d-e2a8-4f20-b213-f47b665d2bbf {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.645657] env[62385]: DEBUG oslo_vmware.api [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205860, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073016} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.645657] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 843.646058] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-591e927b-6bb9-4c68-b362-155aa2860486 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.656483] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-71e9a3c9-f8fb-42f2-a138-6fdd3a07054e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 843.656483] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-71e9a3c9-f8fb-42f2-a138-6fdd3a07054e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 843.656707] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-71e9a3c9-f8fb-42f2-a138-6fdd3a07054e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Deleting the datastore file [datastore1] 55d62bba-ff30-42bf-b1cb-7567988a9361 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 843.657436] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-64ce66d6-b6b4-4b91-b13e-1833edea4c0c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.679466] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] 5a82fc98-e656-4476-b7dc-99466e9d6afd/5a82fc98-e656-4476-b7dc-99466e9d6afd.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 843.682715] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-62dfc14a-b5fc-4a38-9ab2-df23b4d31f28 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.698565] env[62385]: DEBUG oslo_vmware.api [None req-71e9a3c9-f8fb-42f2-a138-6fdd3a07054e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 843.698565] env[62385]: value = "task-1205863" [ 843.698565] env[62385]: _type = "Task" [ 843.698565] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.705922] env[62385]: DEBUG oslo_vmware.api [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 843.705922] env[62385]: value = "task-1205864" [ 843.705922] env[62385]: _type = "Task" [ 843.705922] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.716304] env[62385]: DEBUG oslo_vmware.api [None req-71e9a3c9-f8fb-42f2-a138-6fdd3a07054e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205863, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.719533] env[62385]: DEBUG oslo_vmware.api [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205864, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.740512] env[62385]: DEBUG oslo_vmware.api [None req-a6e09c15-369b-4d28-968e-f207336024dc tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205856, 'name': CloneVM_Task} progress is 94%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.807358] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205861, 'name': CreateVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.822070] env[62385]: DEBUG nova.compute.manager [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Instance disappeared during snapshot {{(pid=62385) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 843.838905] env[62385]: DEBUG nova.compute.manager [None req-9d5340d2-e532-45a6-aa17-54d429720af7 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Image not found during clean up 20957960-bb46-48a4-9b5c-481d6c93295a {{(pid=62385) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4500}} [ 843.866320] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f04fe6a-813a-42a6-bd68-6efab47ac95a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.876795] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e6cf6cf-56fe-4b20-9ae6-c2435b6f4b87 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.913453] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5351542-0126-447e-bae3-08e96de4c0a8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.923385] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-687f7cf0-35cf-4442-9ea0-4714f107f0e1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.939696] env[62385]: DEBUG nova.compute.provider_tree [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 843.967214] env[62385]: DEBUG nova.network.neutron [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 844.065686] env[62385]: DEBUG oslo_concurrency.lockutils [req-8b7342bb-7f6d-49c7-8296-f9cb4d3a5404 req-0dc29ef5-27d7-4976-9bfa-b7294792af9d service nova] Acquiring lock "0d823fcc-cbba-41f6-a47c-2f57d3e79948" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.065791] env[62385]: DEBUG oslo_concurrency.lockutils [req-8b7342bb-7f6d-49c7-8296-f9cb4d3a5404 req-0dc29ef5-27d7-4976-9bfa-b7294792af9d service nova] Acquired lock "0d823fcc-cbba-41f6-a47c-2f57d3e79948" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.066862] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dd4175f-3fd4-4679-b285-baa63d48e976 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.094494] env[62385]: DEBUG oslo_concurrency.lockutils [req-8b7342bb-7f6d-49c7-8296-f9cb4d3a5404 req-0dc29ef5-27d7-4976-9bfa-b7294792af9d service nova] Releasing lock "0d823fcc-cbba-41f6-a47c-2f57d3e79948" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.094809] env[62385]: WARNING nova.compute.manager [req-8b7342bb-7f6d-49c7-8296-f9cb4d3a5404 req-0dc29ef5-27d7-4976-9bfa-b7294792af9d service nova] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Detach interface failed, port_id=51428619-2fcb-4170-97f7-e6b126268421, reason: No device with interface-id 51428619-2fcb-4170-97f7-e6b126268421 exists on VM: nova.exception.NotFound: No device with interface-id 51428619-2fcb-4170-97f7-e6b126268421 exists on VM [ 844.211202] env[62385]: DEBUG oslo_vmware.api [None req-71e9a3c9-f8fb-42f2-a138-6fdd3a07054e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205863, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.376099} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.213157] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-71e9a3c9-f8fb-42f2-a138-6fdd3a07054e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 844.213386] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-71e9a3c9-f8fb-42f2-a138-6fdd3a07054e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 844.213578] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-71e9a3c9-f8fb-42f2-a138-6fdd3a07054e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 844.213765] env[62385]: INFO nova.compute.manager [None req-71e9a3c9-f8fb-42f2-a138-6fdd3a07054e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Took 0.65 seconds to destroy the instance on the hypervisor. [ 844.214022] env[62385]: DEBUG oslo.service.loopingcall [None req-71e9a3c9-f8fb-42f2-a138-6fdd3a07054e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 844.214263] env[62385]: DEBUG nova.compute.manager [-] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 844.214346] env[62385]: DEBUG nova.network.neutron [-] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 844.221743] env[62385]: DEBUG oslo_vmware.api [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205864, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.241429] env[62385]: DEBUG oslo_vmware.api [None req-a6e09c15-369b-4d28-968e-f207336024dc tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205856, 'name': CloneVM_Task, 'duration_secs': 1.51368} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.241698] env[62385]: INFO nova.virt.vmwareapi.vmops [None req-a6e09c15-369b-4d28-968e-f207336024dc tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Created linked-clone VM from snapshot [ 844.242476] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6684c5a0-fc41-4f90-9636-309306f73747 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.250497] env[62385]: DEBUG nova.virt.vmwareapi.images [None req-a6e09c15-369b-4d28-968e-f207336024dc tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Uploading image 496e7b4f-f413-4c90-8a25-d4ab00a67dbf {{(pid=62385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 844.281520] env[62385]: DEBUG oslo_vmware.rw_handles [None req-a6e09c15-369b-4d28-968e-f207336024dc tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 844.281520] env[62385]: value = "vm-261217" [ 844.281520] env[62385]: _type = "VirtualMachine" [ 844.281520] env[62385]: }. {{(pid=62385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 844.281856] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-d1d5a5f4-8bae-4fe2-8a7a-4b8e407e386d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.292703] env[62385]: DEBUG oslo_vmware.rw_handles [None req-a6e09c15-369b-4d28-968e-f207336024dc tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Lease: (returnval){ [ 844.292703] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521f37c2-e475-ca95-a6c8-aa960fabb3d0" [ 844.292703] env[62385]: _type = "HttpNfcLease" [ 844.292703] env[62385]: } obtained for exporting VM: (result){ [ 844.292703] env[62385]: value = "vm-261217" [ 844.292703] env[62385]: _type = "VirtualMachine" [ 844.292703] env[62385]: }. {{(pid=62385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 844.292703] env[62385]: DEBUG oslo_vmware.api [None req-a6e09c15-369b-4d28-968e-f207336024dc tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Waiting for the lease: (returnval){ [ 844.292703] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521f37c2-e475-ca95-a6c8-aa960fabb3d0" [ 844.292703] env[62385]: _type = "HttpNfcLease" [ 844.292703] env[62385]: } to be ready. {{(pid=62385) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 844.307077] env[62385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 844.307077] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521f37c2-e475-ca95-a6c8-aa960fabb3d0" [ 844.307077] env[62385]: _type = "HttpNfcLease" [ 844.307077] env[62385]: } is ready. {{(pid=62385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 844.308355] env[62385]: DEBUG nova.network.neutron [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Updating instance_info_cache with network_info: [{"id": "508791ff-1aa3-43d6-9dd9-1aac3351b688", "address": "fa:16:3e:2a:c1:4d", "network": {"id": "8b2d7a6e-25d4-4a54-ab54-5972612046da", "bridge": "br-int", "label": "tempest-ServersTestJSON-1142134211-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bf8e3c942d3445919cfbe988cca84e90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap508791ff-1a", "ovs_interfaceid": "508791ff-1aa3-43d6-9dd9-1aac3351b688", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.310333] env[62385]: DEBUG oslo_vmware.rw_handles [None req-a6e09c15-369b-4d28-968e-f207336024dc tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 844.310333] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521f37c2-e475-ca95-a6c8-aa960fabb3d0" [ 844.310333] env[62385]: _type = "HttpNfcLease" [ 844.310333] env[62385]: }. {{(pid=62385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 844.311401] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b36822d2-9d60-45e4-9423-56a7703e0b6a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.317989] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205861, 'name': CreateVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.322806] env[62385]: DEBUG oslo_vmware.rw_handles [None req-a6e09c15-369b-4d28-968e-f207336024dc tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52926248-cc62-db81-3e11-08a35f3299a4/disk-0.vmdk from lease info. {{(pid=62385) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 844.322998] env[62385]: DEBUG oslo_vmware.rw_handles [None req-a6e09c15-369b-4d28-968e-f207336024dc tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52926248-cc62-db81-3e11-08a35f3299a4/disk-0.vmdk for reading. {{(pid=62385) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 844.383131] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3e7669c8-b9b4-49bf-88ac-b13f27909a20 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "refresh_cache-0d823fcc-cbba-41f6-a47c-2f57d3e79948" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.383131] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3e7669c8-b9b4-49bf-88ac-b13f27909a20 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquired lock "refresh_cache-0d823fcc-cbba-41f6-a47c-2f57d3e79948" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.383131] env[62385]: DEBUG nova.network.neutron [None req-3e7669c8-b9b4-49bf-88ac-b13f27909a20 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 844.416575] env[62385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-5516fa7b-5cff-41a7-aa98-02845f8301e0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.442984] env[62385]: DEBUG nova.scheduler.client.report [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 844.480873] env[62385]: DEBUG nova.compute.manager [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 844.511575] env[62385]: DEBUG nova.virt.hardware [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 844.511948] env[62385]: DEBUG nova.virt.hardware [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 844.512041] env[62385]: DEBUG nova.virt.hardware [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 844.512644] env[62385]: DEBUG nova.virt.hardware [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 844.512644] env[62385]: DEBUG nova.virt.hardware [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 844.512644] env[62385]: DEBUG nova.virt.hardware [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 844.512800] env[62385]: DEBUG nova.virt.hardware [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 844.512843] env[62385]: DEBUG nova.virt.hardware [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 844.513121] env[62385]: DEBUG nova.virt.hardware [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 844.513375] env[62385]: DEBUG nova.virt.hardware [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 844.513577] env[62385]: DEBUG nova.virt.hardware [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 844.514468] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21984e04-c3eb-47c3-8dbd-d52f6ab3afbb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.522567] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46dfbfe0-9194-4093-ac15-07d5cb57d8d9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.592016] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquiring lock "bd7263b2-b996-4794-946b-2c28215574cb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.592016] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "bd7263b2-b996-4794-946b-2c28215574cb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.721443] env[62385]: DEBUG oslo_vmware.api [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205864, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.808574] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205861, 'name': CreateVM_Task, 'duration_secs': 1.353573} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.808785] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 844.809384] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.809544] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.810129] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 844.810966] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1593217c-f807-443a-9a73-72199d5ae0c7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.815616] env[62385]: DEBUG oslo_concurrency.lockutils [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Releasing lock "refresh_cache-b73a31df-53c8-4550-bf75-5cf3b5aff86c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.815937] env[62385]: DEBUG nova.compute.manager [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Instance network_info: |[{"id": "508791ff-1aa3-43d6-9dd9-1aac3351b688", "address": "fa:16:3e:2a:c1:4d", "network": {"id": "8b2d7a6e-25d4-4a54-ab54-5972612046da", "bridge": "br-int", "label": "tempest-ServersTestJSON-1142134211-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bf8e3c942d3445919cfbe988cca84e90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap508791ff-1a", "ovs_interfaceid": "508791ff-1aa3-43d6-9dd9-1aac3351b688", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 844.816466] env[62385]: DEBUG oslo_vmware.api [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Waiting for the task: (returnval){ [ 844.816466] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52721abb-bf3c-603d-8bf3-063da44bc1ce" [ 844.816466] env[62385]: _type = "Task" [ 844.816466] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.816841] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2a:c1:4d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '508791ff-1aa3-43d6-9dd9-1aac3351b688', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 844.824974] env[62385]: DEBUG oslo.service.loopingcall [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 844.825602] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 844.829435] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2adc3bfe-95d7-4241-ba48-e3440835ff9d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.855505] env[62385]: DEBUG oslo_vmware.api [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52721abb-bf3c-603d-8bf3-063da44bc1ce, 'name': SearchDatastore_Task, 'duration_secs': 0.011325} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.860197] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.860197] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 844.860197] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.860197] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.860197] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 844.860529] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 844.860529] env[62385]: value = "task-1205867" [ 844.860529] env[62385]: _type = "Task" [ 844.860529] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.860759] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5d4dbcce-6421-4d3f-9652-859cc485a812 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.880353] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205867, 'name': CreateVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.882220] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 844.882451] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 844.885057] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a969204-9728-4c3b-8500-c5bcd3b2a0d9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.893399] env[62385]: DEBUG oslo_vmware.api [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Waiting for the task: (returnval){ [ 844.893399] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52820aeb-318d-a3ce-8e85-f8840b6f3849" [ 844.893399] env[62385]: _type = "Task" [ 844.893399] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.905862] env[62385]: DEBUG oslo_vmware.api [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52820aeb-318d-a3ce-8e85-f8840b6f3849, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.954504] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.509s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.955149] env[62385]: DEBUG nova.compute.manager [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 844.960322] env[62385]: DEBUG oslo_concurrency.lockutils [None req-695d2e36-0194-441c-bb9f-b1d1d915536f tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.465s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.960599] env[62385]: DEBUG oslo_concurrency.lockutils [None req-695d2e36-0194-441c-bb9f-b1d1d915536f tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.962786] env[62385]: DEBUG oslo_concurrency.lockutils [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.658s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.964487] env[62385]: INFO nova.compute.claims [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 845.005060] env[62385]: INFO nova.scheduler.client.report [None req-695d2e36-0194-441c-bb9f-b1d1d915536f tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Deleted allocations for instance 2b744879-01d2-440e-8adf-58438b922888 [ 845.072628] env[62385]: DEBUG nova.compute.manager [req-ddc00ae6-6d36-4ccf-b466-b85fadc16a5c req-113c6314-96a6-4233-bf45-2a3fd624357d service nova] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Received event network-changed-508791ff-1aa3-43d6-9dd9-1aac3351b688 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 845.072628] env[62385]: DEBUG nova.compute.manager [req-ddc00ae6-6d36-4ccf-b466-b85fadc16a5c req-113c6314-96a6-4233-bf45-2a3fd624357d service nova] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Refreshing instance network info cache due to event network-changed-508791ff-1aa3-43d6-9dd9-1aac3351b688. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 845.072628] env[62385]: DEBUG oslo_concurrency.lockutils [req-ddc00ae6-6d36-4ccf-b466-b85fadc16a5c req-113c6314-96a6-4233-bf45-2a3fd624357d service nova] Acquiring lock "refresh_cache-b73a31df-53c8-4550-bf75-5cf3b5aff86c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.072628] env[62385]: DEBUG oslo_concurrency.lockutils [req-ddc00ae6-6d36-4ccf-b466-b85fadc16a5c req-113c6314-96a6-4233-bf45-2a3fd624357d service nova] Acquired lock "refresh_cache-b73a31df-53c8-4550-bf75-5cf3b5aff86c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.073215] env[62385]: DEBUG nova.network.neutron [req-ddc00ae6-6d36-4ccf-b466-b85fadc16a5c req-113c6314-96a6-4233-bf45-2a3fd624357d service nova] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Refreshing network info cache for port 508791ff-1aa3-43d6-9dd9-1aac3351b688 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 845.093652] env[62385]: DEBUG nova.compute.manager [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 845.182088] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5dfcbc54-c64f-4274-aed3-ea9558593cc5 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "0d823fcc-cbba-41f6-a47c-2f57d3e79948" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.182415] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5dfcbc54-c64f-4274-aed3-ea9558593cc5 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "0d823fcc-cbba-41f6-a47c-2f57d3e79948" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.182600] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5dfcbc54-c64f-4274-aed3-ea9558593cc5 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "0d823fcc-cbba-41f6-a47c-2f57d3e79948-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.182703] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5dfcbc54-c64f-4274-aed3-ea9558593cc5 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "0d823fcc-cbba-41f6-a47c-2f57d3e79948-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.182877] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5dfcbc54-c64f-4274-aed3-ea9558593cc5 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "0d823fcc-cbba-41f6-a47c-2f57d3e79948-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.189335] env[62385]: INFO nova.compute.manager [None req-5dfcbc54-c64f-4274-aed3-ea9558593cc5 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Terminating instance [ 845.189335] env[62385]: DEBUG nova.compute.manager [None req-5dfcbc54-c64f-4274-aed3-ea9558593cc5 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 845.189335] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-5dfcbc54-c64f-4274-aed3-ea9558593cc5 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 845.190164] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abaeb8b9-f69a-4055-b1ac-b4d206fe9647 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.200378] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dfcbc54-c64f-4274-aed3-ea9558593cc5 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 845.200710] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-880038ab-9842-488e-b22f-d2e1c51351c2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.208440] env[62385]: DEBUG oslo_vmware.api [None req-5dfcbc54-c64f-4274-aed3-ea9558593cc5 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 845.208440] env[62385]: value = "task-1205868" [ 845.208440] env[62385]: _type = "Task" [ 845.208440] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.221044] env[62385]: DEBUG oslo_vmware.api [None req-5dfcbc54-c64f-4274-aed3-ea9558593cc5 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1205868, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.226850] env[62385]: DEBUG oslo_vmware.api [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205864, 'name': ReconfigVM_Task, 'duration_secs': 1.270239} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.227232] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Reconfigured VM instance instance-00000046 to attach disk [datastore2] 5a82fc98-e656-4476-b7dc-99466e9d6afd/5a82fc98-e656-4476-b7dc-99466e9d6afd.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 845.228463] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-219ad590-e513-4a2d-9392-8f94dbb757f5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.234448] env[62385]: DEBUG oslo_vmware.api [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 845.234448] env[62385]: value = "task-1205869" [ 845.234448] env[62385]: _type = "Task" [ 845.234448] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.243774] env[62385]: DEBUG oslo_vmware.api [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205869, 'name': Rename_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.291523] env[62385]: DEBUG nova.network.neutron [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Successfully updated port: 0be859b7-c0bc-435a-908c-49582f2d31da {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 845.380436] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205867, 'name': CreateVM_Task, 'duration_secs': 0.37445} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.381507] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 845.382348] env[62385]: DEBUG oslo_concurrency.lockutils [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.382525] env[62385]: DEBUG oslo_concurrency.lockutils [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.382884] env[62385]: DEBUG oslo_concurrency.lockutils [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 845.383182] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e0792711-2d33-4f29-b0c0-f90fa72de7fd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.388847] env[62385]: DEBUG oslo_vmware.api [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 845.388847] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c6b811-86de-af16-3bba-a28f5e244012" [ 845.388847] env[62385]: _type = "Task" [ 845.388847] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.403311] env[62385]: DEBUG oslo_vmware.api [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c6b811-86de-af16-3bba-a28f5e244012, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.410335] env[62385]: DEBUG oslo_vmware.api [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52820aeb-318d-a3ce-8e85-f8840b6f3849, 'name': SearchDatastore_Task, 'duration_secs': 0.010107} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.411434] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78541d04-84cf-429b-9f7a-b68bfd3024f3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.416982] env[62385]: DEBUG oslo_vmware.api [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Waiting for the task: (returnval){ [ 845.416982] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52cc6c62-09ec-efbf-7279-8478254ece06" [ 845.416982] env[62385]: _type = "Task" [ 845.416982] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.425458] env[62385]: DEBUG oslo_vmware.api [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52cc6c62-09ec-efbf-7279-8478254ece06, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.460999] env[62385]: DEBUG nova.compute.utils [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 845.464304] env[62385]: DEBUG nova.compute.manager [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 845.464501] env[62385]: DEBUG nova.network.neutron [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 845.519152] env[62385]: DEBUG oslo_concurrency.lockutils [None req-695d2e36-0194-441c-bb9f-b1d1d915536f tempest-ServerTagsTestJSON-227488846 tempest-ServerTagsTestJSON-227488846-project-member] Lock "2b744879-01d2-440e-8adf-58438b922888" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.800s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.533102] env[62385]: DEBUG nova.network.neutron [None req-3e7669c8-b9b4-49bf-88ac-b13f27909a20 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Updating instance_info_cache with network_info: [{"id": "ad929652-f520-465a-aeea-1db4e8f36fc3", "address": "fa:16:3e:b6:00:7c", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.130", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad929652-f5", "ovs_interfaceid": "ad929652-f520-465a-aeea-1db4e8f36fc3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.593684] env[62385]: DEBUG nova.policy [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e12425b6f5f84ca7a8ace6cb0fdadb0c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8eaa81326f084441a87ac05796958abb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 845.622860] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.648504] env[62385]: DEBUG nova.network.neutron [-] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.721991] env[62385]: DEBUG oslo_vmware.api [None req-5dfcbc54-c64f-4274-aed3-ea9558593cc5 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1205868, 'name': PowerOffVM_Task, 'duration_secs': 0.217565} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.722319] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dfcbc54-c64f-4274-aed3-ea9558593cc5 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 845.722662] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-5dfcbc54-c64f-4274-aed3-ea9558593cc5 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 845.723010] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c106ab41-dae1-4c08-84c3-ef7d7e80f2d4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.752302] env[62385]: DEBUG oslo_vmware.api [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205869, 'name': Rename_Task, 'duration_secs': 0.159262} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.756554] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 845.756554] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d0f2e60f-a9cd-4bbd-ae85-82f526b89b58 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.761113] env[62385]: DEBUG nova.compute.manager [req-ab0eed6a-630e-4c9e-9218-d3c8b6af27e3 req-2ae41084-e121-4a35-a186-e2409b7fb332 service nova] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Received event network-vif-deleted-5f63776d-8e86-4723-b873-9a8162076910 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 845.765500] env[62385]: DEBUG oslo_vmware.api [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 845.765500] env[62385]: value = "task-1205871" [ 845.765500] env[62385]: _type = "Task" [ 845.765500] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.774473] env[62385]: DEBUG oslo_vmware.api [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205871, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.790228] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-5dfcbc54-c64f-4274-aed3-ea9558593cc5 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 845.791886] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-5dfcbc54-c64f-4274-aed3-ea9558593cc5 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 845.791886] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-5dfcbc54-c64f-4274-aed3-ea9558593cc5 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Deleting the datastore file [datastore1] 0d823fcc-cbba-41f6-a47c-2f57d3e79948 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 845.791886] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1a215323-7e81-416b-ad04-88705280c4f1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.795050] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Acquiring lock "refresh_cache-2a579d13-5372-4340-b7b3-cc02c1912624" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.795427] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Acquired lock "refresh_cache-2a579d13-5372-4340-b7b3-cc02c1912624" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.795653] env[62385]: DEBUG nova.network.neutron [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 845.803476] env[62385]: DEBUG oslo_vmware.api [None req-5dfcbc54-c64f-4274-aed3-ea9558593cc5 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 845.803476] env[62385]: value = "task-1205872" [ 845.803476] env[62385]: _type = "Task" [ 845.803476] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.817272] env[62385]: DEBUG oslo_vmware.api [None req-5dfcbc54-c64f-4274-aed3-ea9558593cc5 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1205872, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.901852] env[62385]: DEBUG oslo_vmware.api [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c6b811-86de-af16-3bba-a28f5e244012, 'name': SearchDatastore_Task, 'duration_secs': 0.010142} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.904609] env[62385]: DEBUG oslo_concurrency.lockutils [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.904609] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 845.904609] env[62385]: DEBUG oslo_concurrency.lockutils [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.904609] env[62385]: DEBUG oslo_concurrency.lockutils [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.904609] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 845.904609] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-406bea5b-2bd2-4002-8787-0559080e6e4d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.917812] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 845.917812] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 845.917812] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6f281512-3bb2-4751-b8ee-58efbe6daa28 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.928613] env[62385]: DEBUG oslo_vmware.api [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 845.928613] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52bda4f6-3c6d-c279-28a8-59a5c50f1dd6" [ 845.928613] env[62385]: _type = "Task" [ 845.928613] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.932021] env[62385]: DEBUG oslo_vmware.api [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52cc6c62-09ec-efbf-7279-8478254ece06, 'name': SearchDatastore_Task, 'duration_secs': 0.018755} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.936595] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.936925] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 7f2fe830-e16d-4684-91a4-aab219468e77/7f2fe830-e16d-4684-91a4-aab219468e77.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 845.937272] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bafee50f-49d0-47d8-a47e-47964303641c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.946680] env[62385]: DEBUG oslo_vmware.api [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52bda4f6-3c6d-c279-28a8-59a5c50f1dd6, 'name': SearchDatastore_Task, 'duration_secs': 0.013378} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.949364] env[62385]: DEBUG oslo_vmware.api [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Waiting for the task: (returnval){ [ 845.949364] env[62385]: value = "task-1205873" [ 845.949364] env[62385]: _type = "Task" [ 845.949364] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.949364] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-790cb60d-1f60-406e-8be9-aa254ad326ad {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.958622] env[62385]: DEBUG oslo_vmware.api [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 845.958622] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]527aee1f-85b4-7ae1-7de9-42486025fe89" [ 845.958622] env[62385]: _type = "Task" [ 845.958622] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.965458] env[62385]: DEBUG nova.compute.manager [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 845.968191] env[62385]: DEBUG oslo_vmware.api [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': task-1205873, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.981628] env[62385]: DEBUG oslo_vmware.api [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]527aee1f-85b4-7ae1-7de9-42486025fe89, 'name': SearchDatastore_Task, 'duration_secs': 0.011406} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.981825] env[62385]: DEBUG oslo_concurrency.lockutils [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.982248] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] b73a31df-53c8-4550-bf75-5cf3b5aff86c/b73a31df-53c8-4550-bf75-5cf3b5aff86c.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 845.982663] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3af308bf-a7b4-48f3-a900-abde2a0e74ae {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.991760] env[62385]: DEBUG oslo_vmware.api [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 845.991760] env[62385]: value = "task-1205874" [ 845.991760] env[62385]: _type = "Task" [ 845.991760] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.000680] env[62385]: DEBUG oslo_vmware.api [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1205874, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.013037] env[62385]: DEBUG nova.network.neutron [req-ddc00ae6-6d36-4ccf-b466-b85fadc16a5c req-113c6314-96a6-4233-bf45-2a3fd624357d service nova] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Updated VIF entry in instance network info cache for port 508791ff-1aa3-43d6-9dd9-1aac3351b688. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 846.014019] env[62385]: DEBUG nova.network.neutron [req-ddc00ae6-6d36-4ccf-b466-b85fadc16a5c req-113c6314-96a6-4233-bf45-2a3fd624357d service nova] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Updating instance_info_cache with network_info: [{"id": "508791ff-1aa3-43d6-9dd9-1aac3351b688", "address": "fa:16:3e:2a:c1:4d", "network": {"id": "8b2d7a6e-25d4-4a54-ab54-5972612046da", "bridge": "br-int", "label": "tempest-ServersTestJSON-1142134211-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bf8e3c942d3445919cfbe988cca84e90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap508791ff-1a", "ovs_interfaceid": "508791ff-1aa3-43d6-9dd9-1aac3351b688", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.039699] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3e7669c8-b9b4-49bf-88ac-b13f27909a20 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Releasing lock "refresh_cache-0d823fcc-cbba-41f6-a47c-2f57d3e79948" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.151629] env[62385]: INFO nova.compute.manager [-] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Took 1.94 seconds to deallocate network for instance. [ 846.278541] env[62385]: DEBUG oslo_vmware.api [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205871, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.318589] env[62385]: DEBUG oslo_vmware.api [None req-5dfcbc54-c64f-4274-aed3-ea9558593cc5 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1205872, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.18641} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.318589] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-5dfcbc54-c64f-4274-aed3-ea9558593cc5 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 846.318589] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-5dfcbc54-c64f-4274-aed3-ea9558593cc5 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 846.318745] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-5dfcbc54-c64f-4274-aed3-ea9558593cc5 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 846.319166] env[62385]: INFO nova.compute.manager [None req-5dfcbc54-c64f-4274-aed3-ea9558593cc5 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Took 1.13 seconds to destroy the instance on the hypervisor. [ 846.319515] env[62385]: DEBUG oslo.service.loopingcall [None req-5dfcbc54-c64f-4274-aed3-ea9558593cc5 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 846.319710] env[62385]: DEBUG nova.compute.manager [-] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 846.319807] env[62385]: DEBUG nova.network.neutron [-] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 846.418226] env[62385]: DEBUG nova.network.neutron [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 846.443032] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0fbb98a-8138-49e1-973b-f53861840346 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.451531] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4203a88b-b93d-4ef6-986b-1cac4b754b44 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.473788] env[62385]: DEBUG oslo_vmware.api [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': task-1205873, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.506770] env[62385]: DEBUG nova.network.neutron [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Successfully created port: 823867e4-d6b8-42d1-8597-3c27f6190ab1 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 846.521025] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e200f40-0f58-4888-aa70-818ef83c46fb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.526167] env[62385]: DEBUG oslo_concurrency.lockutils [req-ddc00ae6-6d36-4ccf-b466-b85fadc16a5c req-113c6314-96a6-4233-bf45-2a3fd624357d service nova] Releasing lock "refresh_cache-b73a31df-53c8-4550-bf75-5cf3b5aff86c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.536067] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a22ed8bc-0296-4006-98dd-6b7b1dda721b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.540515] env[62385]: DEBUG oslo_vmware.api [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1205874, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.551541] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3e7669c8-b9b4-49bf-88ac-b13f27909a20 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "interface-0d823fcc-cbba-41f6-a47c-2f57d3e79948-51428619-2fcb-4170-97f7-e6b126268421" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.195s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.552941] env[62385]: DEBUG nova.compute.provider_tree [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Updating inventory in ProviderTree for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 846.665203] env[62385]: DEBUG oslo_concurrency.lockutils [None req-71e9a3c9-f8fb-42f2-a138-6fdd3a07054e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.777207] env[62385]: DEBUG oslo_vmware.api [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205871, 'name': PowerOnVM_Task, 'duration_secs': 0.512242} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.777523] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 846.777768] env[62385]: DEBUG nova.compute.manager [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 846.778981] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e31e7a9-9783-4901-bb75-e5229d02c49d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.969689] env[62385]: DEBUG oslo_vmware.api [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': task-1205873, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.628154} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.969962] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 7f2fe830-e16d-4684-91a4-aab219468e77/7f2fe830-e16d-4684-91a4-aab219468e77.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 846.970215] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 846.970470] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-25f8f3af-4e6c-4d14-8c7a-51f308509664 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.977296] env[62385]: DEBUG oslo_vmware.api [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Waiting for the task: (returnval){ [ 846.977296] env[62385]: value = "task-1205875" [ 846.977296] env[62385]: _type = "Task" [ 846.977296] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.985900] env[62385]: DEBUG oslo_vmware.api [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': task-1205875, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.015822] env[62385]: DEBUG nova.compute.manager [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 847.018343] env[62385]: DEBUG oslo_vmware.api [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1205874, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.645696} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.021565] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] b73a31df-53c8-4550-bf75-5cf3b5aff86c/b73a31df-53c8-4550-bf75-5cf3b5aff86c.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 847.021565] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 847.021565] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5dc71bcb-d3b8-4127-87f9-7bee8b2e13c8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.028162] env[62385]: DEBUG oslo_vmware.api [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 847.028162] env[62385]: value = "task-1205876" [ 847.028162] env[62385]: _type = "Task" [ 847.028162] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.032564] env[62385]: DEBUG nova.network.neutron [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Updating instance_info_cache with network_info: [{"id": "0be859b7-c0bc-435a-908c-49582f2d31da", "address": "fa:16:3e:6d:0d:d0", "network": {"id": "3325f6ee-8d27-49f8-8959-af54a4bb2230", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1230956501-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "212a132650584a4d866dc95004faa391", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0be859b7-c0", "ovs_interfaceid": "0be859b7-c0bc-435a-908c-49582f2d31da", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.039538] env[62385]: DEBUG oslo_vmware.api [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1205876, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.073493] env[62385]: DEBUG nova.virt.hardware [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 847.074041] env[62385]: DEBUG nova.virt.hardware [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 847.074906] env[62385]: DEBUG nova.virt.hardware [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 847.074906] env[62385]: DEBUG nova.virt.hardware [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 847.074906] env[62385]: DEBUG nova.virt.hardware [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 847.074906] env[62385]: DEBUG nova.virt.hardware [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 847.075141] env[62385]: DEBUG nova.virt.hardware [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 847.075426] env[62385]: DEBUG nova.virt.hardware [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 847.075503] env[62385]: DEBUG nova.virt.hardware [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 847.075721] env[62385]: DEBUG nova.virt.hardware [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 847.075915] env[62385]: DEBUG nova.virt.hardware [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 847.079298] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c27a589-ca6e-4715-9475-db5890482de2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.082026] env[62385]: ERROR nova.scheduler.client.report [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] [req-318978ae-b7de-45ac-9f17-e3ed199079bf] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-318978ae-b7de-45ac-9f17-e3ed199079bf"}]} [ 847.089650] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97ac0a8d-7358-4ac7-bf7c-67dd64aa2bdb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.106864] env[62385]: DEBUG nova.scheduler.client.report [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Refreshing inventories for resource provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 847.131020] env[62385]: DEBUG nova.scheduler.client.report [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Updating ProviderTree inventory for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 847.131325] env[62385]: DEBUG nova.compute.provider_tree [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Updating inventory in ProviderTree for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 847.144202] env[62385]: DEBUG nova.scheduler.client.report [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Refreshing aggregate associations for resource provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a, aggregates: None {{(pid=62385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 847.167844] env[62385]: DEBUG nova.scheduler.client.report [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Refreshing trait associations for resource provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=62385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 847.305451] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.359469] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-2aecba33-8842-48e6-bc38-81bd5b2a0bec tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Volume attach. Driver type: vmdk {{(pid=62385) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 847.359794] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-2aecba33-8842-48e6-bc38-81bd5b2a0bec tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261218', 'volume_id': 'ddb60eb1-ec31-4439-b8c9-0c53408d8c97', 'name': 'volume-ddb60eb1-ec31-4439-b8c9-0c53408d8c97', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd646c9a3-0ed3-4e5b-9fc9-886a1a507f1b', 'attached_at': '', 'detached_at': '', 'volume_id': 'ddb60eb1-ec31-4439-b8c9-0c53408d8c97', 'serial': 'ddb60eb1-ec31-4439-b8c9-0c53408d8c97'} {{(pid=62385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 847.360750] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01e587f9-a7fa-40a0-8965-e1a5f199b739 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.382902] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af1d500d-69d4-4dd3-8422-b87784e3d419 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.413331] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-2aecba33-8842-48e6-bc38-81bd5b2a0bec tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] volume-ddb60eb1-ec31-4439-b8c9-0c53408d8c97/volume-ddb60eb1-ec31-4439-b8c9-0c53408d8c97.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 847.416873] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8982f068-4d66-41aa-8692-0e7c5c22ac51 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.439020] env[62385]: DEBUG oslo_vmware.api [None req-2aecba33-8842-48e6-bc38-81bd5b2a0bec tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for the task: (returnval){ [ 847.439020] env[62385]: value = "task-1205877" [ 847.439020] env[62385]: _type = "Task" [ 847.439020] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.448304] env[62385]: DEBUG oslo_vmware.api [None req-2aecba33-8842-48e6-bc38-81bd5b2a0bec tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1205877, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.489768] env[62385]: DEBUG oslo_vmware.api [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': task-1205875, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068926} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.493073] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 847.494356] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5af3cdc4-1a68-446a-bc3b-fa37b8b3ff28 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.518248] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Reconfiguring VM instance instance-00000047 to attach disk [datastore1] 7f2fe830-e16d-4684-91a4-aab219468e77/7f2fe830-e16d-4684-91a4-aab219468e77.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 847.521482] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-92fc9336-269e-470a-bd33-781f057a1f0d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.538793] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Releasing lock "refresh_cache-2a579d13-5372-4340-b7b3-cc02c1912624" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.538793] env[62385]: DEBUG nova.compute.manager [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Instance network_info: |[{"id": "0be859b7-c0bc-435a-908c-49582f2d31da", "address": "fa:16:3e:6d:0d:d0", "network": {"id": "3325f6ee-8d27-49f8-8959-af54a4bb2230", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1230956501-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "212a132650584a4d866dc95004faa391", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0be859b7-c0", "ovs_interfaceid": "0be859b7-c0bc-435a-908c-49582f2d31da", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 847.542804] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6d:0d:d0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'db68bd64-5b56-49af-a075-13dcf85cb2e0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0be859b7-c0bc-435a-908c-49582f2d31da', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 847.550533] env[62385]: DEBUG oslo.service.loopingcall [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 847.551340] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 847.551937] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a8d48296-ef5d-40b1-9626-663c255d6cf3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.576235] env[62385]: DEBUG oslo_vmware.api [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1205876, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.094142} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.576788] env[62385]: DEBUG oslo_vmware.api [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Waiting for the task: (returnval){ [ 847.576788] env[62385]: value = "task-1205878" [ 847.576788] env[62385]: _type = "Task" [ 847.576788] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.577773] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 847.579017] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc6407df-2bea-470e-bb01-035b0b8d1796 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.586728] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 847.586728] env[62385]: value = "task-1205879" [ 847.586728] env[62385]: _type = "Task" [ 847.586728] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.603512] env[62385]: DEBUG oslo_vmware.api [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': task-1205878, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.613038] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] b73a31df-53c8-4550-bf75-5cf3b5aff86c/b73a31df-53c8-4550-bf75-5cf3b5aff86c.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 847.616565] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8dbfde95-cba6-4629-a189-2c4623c049ed {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.639158] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205879, 'name': CreateVM_Task} progress is 15%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.644443] env[62385]: DEBUG oslo_vmware.api [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 847.644443] env[62385]: value = "task-1205880" [ 847.644443] env[62385]: _type = "Task" [ 847.644443] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.658375] env[62385]: DEBUG oslo_vmware.api [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1205880, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.682875] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19fdedfa-5047-4657-bd59-b8550ed0e2da {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.691608] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4271a69-7169-45a4-8560-b7f2f72896a0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.725091] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b2bed02-e75d-4601-9b12-22427d78959f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.733743] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d1b4fb0-dcad-4d6b-9869-7403e07154b4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.748183] env[62385]: DEBUG nova.compute.provider_tree [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 847.950928] env[62385]: DEBUG oslo_vmware.api [None req-2aecba33-8842-48e6-bc38-81bd5b2a0bec tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1205877, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.024142] env[62385]: DEBUG nova.compute.manager [req-c63f425c-04ed-4baa-ae84-8e287dede5fa req-6e6ae392-6be3-4dcc-b0e1-9d5990887758 service nova] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Received event network-vif-plugged-0be859b7-c0bc-435a-908c-49582f2d31da {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 848.024142] env[62385]: DEBUG oslo_concurrency.lockutils [req-c63f425c-04ed-4baa-ae84-8e287dede5fa req-6e6ae392-6be3-4dcc-b0e1-9d5990887758 service nova] Acquiring lock "2a579d13-5372-4340-b7b3-cc02c1912624-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.024378] env[62385]: DEBUG oslo_concurrency.lockutils [req-c63f425c-04ed-4baa-ae84-8e287dede5fa req-6e6ae392-6be3-4dcc-b0e1-9d5990887758 service nova] Lock "2a579d13-5372-4340-b7b3-cc02c1912624-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.024568] env[62385]: DEBUG oslo_concurrency.lockutils [req-c63f425c-04ed-4baa-ae84-8e287dede5fa req-6e6ae392-6be3-4dcc-b0e1-9d5990887758 service nova] Lock "2a579d13-5372-4340-b7b3-cc02c1912624-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.024740] env[62385]: DEBUG nova.compute.manager [req-c63f425c-04ed-4baa-ae84-8e287dede5fa req-6e6ae392-6be3-4dcc-b0e1-9d5990887758 service nova] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] No waiting events found dispatching network-vif-plugged-0be859b7-c0bc-435a-908c-49582f2d31da {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 848.025021] env[62385]: WARNING nova.compute.manager [req-c63f425c-04ed-4baa-ae84-8e287dede5fa req-6e6ae392-6be3-4dcc-b0e1-9d5990887758 service nova] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Received unexpected event network-vif-plugged-0be859b7-c0bc-435a-908c-49582f2d31da for instance with vm_state building and task_state spawning. [ 848.025103] env[62385]: DEBUG nova.compute.manager [req-c63f425c-04ed-4baa-ae84-8e287dede5fa req-6e6ae392-6be3-4dcc-b0e1-9d5990887758 service nova] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Received event network-changed-0be859b7-c0bc-435a-908c-49582f2d31da {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 848.025286] env[62385]: DEBUG nova.compute.manager [req-c63f425c-04ed-4baa-ae84-8e287dede5fa req-6e6ae392-6be3-4dcc-b0e1-9d5990887758 service nova] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Refreshing instance network info cache due to event network-changed-0be859b7-c0bc-435a-908c-49582f2d31da. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 848.025487] env[62385]: DEBUG oslo_concurrency.lockutils [req-c63f425c-04ed-4baa-ae84-8e287dede5fa req-6e6ae392-6be3-4dcc-b0e1-9d5990887758 service nova] Acquiring lock "refresh_cache-2a579d13-5372-4340-b7b3-cc02c1912624" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.025627] env[62385]: DEBUG oslo_concurrency.lockutils [req-c63f425c-04ed-4baa-ae84-8e287dede5fa req-6e6ae392-6be3-4dcc-b0e1-9d5990887758 service nova] Acquired lock "refresh_cache-2a579d13-5372-4340-b7b3-cc02c1912624" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.025785] env[62385]: DEBUG nova.network.neutron [req-c63f425c-04ed-4baa-ae84-8e287dede5fa req-6e6ae392-6be3-4dcc-b0e1-9d5990887758 service nova] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Refreshing network info cache for port 0be859b7-c0bc-435a-908c-49582f2d31da {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 848.092847] env[62385]: DEBUG oslo_vmware.api [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': task-1205878, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.108305] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205879, 'name': CreateVM_Task} progress is 25%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.156885] env[62385]: DEBUG oslo_vmware.api [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1205880, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.253192] env[62385]: DEBUG nova.scheduler.client.report [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 848.450202] env[62385]: DEBUG oslo_vmware.api [None req-2aecba33-8842-48e6-bc38-81bd5b2a0bec tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1205877, 'name': ReconfigVM_Task, 'duration_secs': 0.921734} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.450202] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-2aecba33-8842-48e6-bc38-81bd5b2a0bec tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Reconfigured VM instance instance-0000003e to attach disk [datastore2] volume-ddb60eb1-ec31-4439-b8c9-0c53408d8c97/volume-ddb60eb1-ec31-4439-b8c9-0c53408d8c97.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 848.455346] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-16dc1f15-7fd3-43ae-a764-32b4fe56c8a9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.474691] env[62385]: DEBUG oslo_vmware.api [None req-2aecba33-8842-48e6-bc38-81bd5b2a0bec tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for the task: (returnval){ [ 848.474691] env[62385]: value = "task-1205881" [ 848.474691] env[62385]: _type = "Task" [ 848.474691] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.486118] env[62385]: DEBUG oslo_vmware.api [None req-2aecba33-8842-48e6-bc38-81bd5b2a0bec tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1205881, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.566627] env[62385]: DEBUG nova.network.neutron [-] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.593459] env[62385]: DEBUG oslo_vmware.api [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': task-1205878, 'name': ReconfigVM_Task, 'duration_secs': 0.616871} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.593459] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Reconfigured VM instance instance-00000047 to attach disk [datastore1] 7f2fe830-e16d-4684-91a4-aab219468e77/7f2fe830-e16d-4684-91a4-aab219468e77.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 848.597162] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f8299cef-8df6-4cac-8351-67d83846a941 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.608103] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205879, 'name': CreateVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.610218] env[62385]: DEBUG oslo_vmware.api [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Waiting for the task: (returnval){ [ 848.610218] env[62385]: value = "task-1205882" [ 848.610218] env[62385]: _type = "Task" [ 848.610218] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.620685] env[62385]: DEBUG oslo_vmware.api [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': task-1205882, 'name': Rename_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.663118] env[62385]: DEBUG oslo_vmware.api [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1205880, 'name': ReconfigVM_Task, 'duration_secs': 1.006477} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.663118] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Reconfigured VM instance instance-00000048 to attach disk [datastore2] b73a31df-53c8-4550-bf75-5cf3b5aff86c/b73a31df-53c8-4550-bf75-5cf3b5aff86c.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 848.663515] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b5475755-ee6d-4543-9a7e-220ed16bafba {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.672965] env[62385]: DEBUG oslo_vmware.api [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 848.672965] env[62385]: value = "task-1205883" [ 848.672965] env[62385]: _type = "Task" [ 848.672965] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.688030] env[62385]: DEBUG oslo_vmware.api [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1205883, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.764667] env[62385]: DEBUG oslo_concurrency.lockutils [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.802s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.766253] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.471s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.768245] env[62385]: INFO nova.compute.claims [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 848.784207] env[62385]: DEBUG nova.compute.manager [req-b1fee78e-054f-4c41-ab15-6e4a671686a7 req-13c5ea6a-63fc-40f5-a3f8-e879b122ead3 service nova] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Received event network-vif-deleted-ad929652-f520-465a-aeea-1db4e8f36fc3 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 848.942032] env[62385]: DEBUG oslo_concurrency.lockutils [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "d12ffea4-2275-4974-a5c6-b7fa365865b4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.942324] env[62385]: DEBUG oslo_concurrency.lockutils [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "d12ffea4-2275-4974-a5c6-b7fa365865b4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.990109] env[62385]: DEBUG oslo_vmware.api [None req-2aecba33-8842-48e6-bc38-81bd5b2a0bec tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1205881, 'name': ReconfigVM_Task, 'duration_secs': 0.231198} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.990109] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-2aecba33-8842-48e6-bc38-81bd5b2a0bec tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261218', 'volume_id': 'ddb60eb1-ec31-4439-b8c9-0c53408d8c97', 'name': 'volume-ddb60eb1-ec31-4439-b8c9-0c53408d8c97', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd646c9a3-0ed3-4e5b-9fc9-886a1a507f1b', 'attached_at': '', 'detached_at': '', 'volume_id': 'ddb60eb1-ec31-4439-b8c9-0c53408d8c97', 'serial': 'ddb60eb1-ec31-4439-b8c9-0c53408d8c97'} {{(pid=62385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 848.994519] env[62385]: DEBUG nova.network.neutron [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Successfully updated port: 823867e4-d6b8-42d1-8597-3c27f6190ab1 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 849.069996] env[62385]: INFO nova.compute.manager [-] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Took 2.75 seconds to deallocate network for instance. [ 849.107707] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205879, 'name': CreateVM_Task, 'duration_secs': 1.133929} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.107707] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 849.107707] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.107707] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.107707] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 849.107707] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f11ffd92-5aa6-4974-9c09-8edd6239bce9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.118168] env[62385]: DEBUG oslo_vmware.api [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for the task: (returnval){ [ 849.118168] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52b9d4a1-12be-ba68-e5ca-2b2595b26b37" [ 849.118168] env[62385]: _type = "Task" [ 849.118168] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.124600] env[62385]: DEBUG nova.network.neutron [req-c63f425c-04ed-4baa-ae84-8e287dede5fa req-6e6ae392-6be3-4dcc-b0e1-9d5990887758 service nova] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Updated VIF entry in instance network info cache for port 0be859b7-c0bc-435a-908c-49582f2d31da. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 849.125120] env[62385]: DEBUG nova.network.neutron [req-c63f425c-04ed-4baa-ae84-8e287dede5fa req-6e6ae392-6be3-4dcc-b0e1-9d5990887758 service nova] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Updating instance_info_cache with network_info: [{"id": "0be859b7-c0bc-435a-908c-49582f2d31da", "address": "fa:16:3e:6d:0d:d0", "network": {"id": "3325f6ee-8d27-49f8-8959-af54a4bb2230", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1230956501-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "212a132650584a4d866dc95004faa391", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0be859b7-c0", "ovs_interfaceid": "0be859b7-c0bc-435a-908c-49582f2d31da", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.130937] env[62385]: DEBUG oslo_vmware.api [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': task-1205882, 'name': Rename_Task, 'duration_secs': 0.197806} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.133613] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 849.134431] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-09ce1964-e7a5-4b9f-9ad3-8a537036986d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.142858] env[62385]: DEBUG oslo_vmware.api [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52b9d4a1-12be-ba68-e5ca-2b2595b26b37, 'name': SearchDatastore_Task, 'duration_secs': 0.013464} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.143756] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.145278] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 849.145278] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.145278] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.145278] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 849.145762] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-85512af0-2195-4b0c-930f-e6f7980888b9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.150996] env[62385]: DEBUG oslo_vmware.api [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Waiting for the task: (returnval){ [ 849.150996] env[62385]: value = "task-1205884" [ 849.150996] env[62385]: _type = "Task" [ 849.150996] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.158631] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 849.158874] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 849.163393] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b0788f2-f282-4bce-a430-6846cf8b4b94 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.167100] env[62385]: DEBUG oslo_vmware.api [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': task-1205884, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.172595] env[62385]: DEBUG oslo_vmware.api [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for the task: (returnval){ [ 849.172595] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52bde263-e909-fec2-f6eb-818c3f22e04b" [ 849.172595] env[62385]: _type = "Task" [ 849.172595] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.194339] env[62385]: DEBUG oslo_vmware.api [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52bde263-e909-fec2-f6eb-818c3f22e04b, 'name': SearchDatastore_Task, 'duration_secs': 0.015003} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.194598] env[62385]: DEBUG oslo_vmware.api [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1205883, 'name': Rename_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.195473] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4bda564-2876-43b2-8c80-51597131019b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.201864] env[62385]: DEBUG oslo_vmware.api [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for the task: (returnval){ [ 849.201864] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52b5aa4d-1a96-3062-52ab-191dc36a84ad" [ 849.201864] env[62385]: _type = "Task" [ 849.201864] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.211599] env[62385]: DEBUG oslo_vmware.api [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52b5aa4d-1a96-3062-52ab-191dc36a84ad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.275286] env[62385]: DEBUG oslo_concurrency.lockutils [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Acquiring lock "a181222c-267d-4d5a-b9dc-c1c7b56a3bc8" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.275286] env[62385]: DEBUG oslo_concurrency.lockutils [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Lock "a181222c-267d-4d5a-b9dc-c1c7b56a3bc8" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.447945] env[62385]: DEBUG nova.compute.manager [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 849.503174] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Acquiring lock "refresh_cache-5e5f43e8-ce45-45c7-a244-bfa724ea5e3c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.503424] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Acquired lock "refresh_cache-5e5f43e8-ce45-45c7-a244-bfa724ea5e3c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.503477] env[62385]: DEBUG nova.network.neutron [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 849.580279] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5dfcbc54-c64f-4274-aed3-ea9558593cc5 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.632735] env[62385]: DEBUG oslo_concurrency.lockutils [req-c63f425c-04ed-4baa-ae84-8e287dede5fa req-6e6ae392-6be3-4dcc-b0e1-9d5990887758 service nova] Releasing lock "refresh_cache-2a579d13-5372-4340-b7b3-cc02c1912624" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.662346] env[62385]: DEBUG oslo_vmware.api [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': task-1205884, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.689852] env[62385]: DEBUG oslo_vmware.api [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1205883, 'name': Rename_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.717117] env[62385]: DEBUG oslo_vmware.api [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52b5aa4d-1a96-3062-52ab-191dc36a84ad, 'name': SearchDatastore_Task, 'duration_secs': 0.013774} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.717405] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.717692] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 2a579d13-5372-4340-b7b3-cc02c1912624/2a579d13-5372-4340-b7b3-cc02c1912624.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 849.717998] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-eddc457e-ef95-4f21-af0b-b5f670f86d08 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.728258] env[62385]: DEBUG oslo_vmware.api [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for the task: (returnval){ [ 849.728258] env[62385]: value = "task-1205885" [ 849.728258] env[62385]: _type = "Task" [ 849.728258] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.739995] env[62385]: DEBUG oslo_vmware.api [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205885, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.780479] env[62385]: DEBUG oslo_concurrency.lockutils [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Lock "a181222c-267d-4d5a-b9dc-c1c7b56a3bc8" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.506s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.781433] env[62385]: DEBUG nova.compute.manager [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 849.985824] env[62385]: DEBUG oslo_concurrency.lockutils [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.042986] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f2cd818c-6b68-4f26-b95e-028548504a6d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "5a82fc98-e656-4476-b7dc-99466e9d6afd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.042986] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f2cd818c-6b68-4f26-b95e-028548504a6d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "5a82fc98-e656-4476-b7dc-99466e9d6afd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.042986] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f2cd818c-6b68-4f26-b95e-028548504a6d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "5a82fc98-e656-4476-b7dc-99466e9d6afd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.043370] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f2cd818c-6b68-4f26-b95e-028548504a6d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "5a82fc98-e656-4476-b7dc-99466e9d6afd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.043370] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f2cd818c-6b68-4f26-b95e-028548504a6d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "5a82fc98-e656-4476-b7dc-99466e9d6afd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.047111] env[62385]: INFO nova.compute.manager [None req-f2cd818c-6b68-4f26-b95e-028548504a6d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Terminating instance [ 850.049617] env[62385]: DEBUG nova.compute.manager [None req-f2cd818c-6b68-4f26-b95e-028548504a6d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 850.049887] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-f2cd818c-6b68-4f26-b95e-028548504a6d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 850.050792] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8ec71ec-d8a1-4f54-a750-99a4a93f0427 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.066667] env[62385]: DEBUG nova.objects.instance [None req-2aecba33-8842-48e6-bc38-81bd5b2a0bec tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lazy-loading 'flavor' on Instance uuid d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 850.067504] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2cd818c-6b68-4f26-b95e-028548504a6d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 850.067853] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cb034a74-aa5a-4363-8cd8-e62487a7ab54 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.076712] env[62385]: DEBUG oslo_vmware.api [None req-f2cd818c-6b68-4f26-b95e-028548504a6d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 850.076712] env[62385]: value = "task-1205886" [ 850.076712] env[62385]: _type = "Task" [ 850.076712] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.093608] env[62385]: DEBUG oslo_vmware.api [None req-f2cd818c-6b68-4f26-b95e-028548504a6d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205886, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.108717] env[62385]: DEBUG nova.network.neutron [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 850.182481] env[62385]: DEBUG oslo_vmware.api [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': task-1205884, 'name': PowerOnVM_Task, 'duration_secs': 0.729154} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.192107] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 850.192628] env[62385]: DEBUG nova.compute.manager [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 850.193724] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9ae12d6-06f4-451d-9d52-782a6cac2df4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.205036] env[62385]: DEBUG oslo_vmware.api [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1205883, 'name': Rename_Task, 'duration_secs': 1.218584} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.211553] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 850.213294] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fe40669a-12d4-492f-aa17-ef4cdc12c1cc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.222532] env[62385]: DEBUG oslo_vmware.api [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 850.222532] env[62385]: value = "task-1205887" [ 850.222532] env[62385]: _type = "Task" [ 850.222532] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.236014] env[62385]: DEBUG oslo_vmware.api [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1205887, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.256138] env[62385]: DEBUG oslo_vmware.api [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205885, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.296051] env[62385]: DEBUG nova.compute.utils [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 850.296051] env[62385]: DEBUG nova.compute.manager [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 850.296051] env[62385]: DEBUG nova.network.neutron [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 850.346019] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce211563-01a1-4ec2-8888-b9eeed0f1f54 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.354541] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd057a7a-835a-4e89-8ac6-1b09339e063b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.393703] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6ede592-1995-4667-9f8d-2f9209433ba0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.401744] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29e9dec1-9c0e-4662-ac27-6ddb067eb412 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.416098] env[62385]: DEBUG nova.compute.provider_tree [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 850.420806] env[62385]: DEBUG nova.policy [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '19c807ab7e8e471aa98fbb4b934cc0c2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c556f4359bd44d2f868a796249b25694', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 850.514025] env[62385]: DEBUG nova.network.neutron [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Updating instance_info_cache with network_info: [{"id": "823867e4-d6b8-42d1-8597-3c27f6190ab1", "address": "fa:16:3e:56:b4:61", "network": {"id": "b563b70c-d3f1-4f82-97e9-defaf3f9d14b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-55206131-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "8eaa81326f084441a87ac05796958abb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa", "external-id": "nsx-vlan-transportzone-132", "segmentation_id": 132, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap823867e4-d6", "ovs_interfaceid": "823867e4-d6b8-42d1-8597-3c27f6190ab1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.575144] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2aecba33-8842-48e6-bc38-81bd5b2a0bec tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.364s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.600019] env[62385]: DEBUG oslo_vmware.api [None req-f2cd818c-6b68-4f26-b95e-028548504a6d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205886, 'name': PowerOffVM_Task, 'duration_secs': 0.282268} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.600019] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2cd818c-6b68-4f26-b95e-028548504a6d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 850.600019] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-f2cd818c-6b68-4f26-b95e-028548504a6d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 850.600019] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ade26ae3-ff64-4cda-a200-db9f79930f67 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.686276] env[62385]: DEBUG nova.compute.manager [req-ffd3d156-f053-48ee-abef-6c94e28c392b req-ef8aa07e-b402-4f7e-9444-aace1ddb83b8 service nova] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Received event network-vif-plugged-823867e4-d6b8-42d1-8597-3c27f6190ab1 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 850.686589] env[62385]: DEBUG oslo_concurrency.lockutils [req-ffd3d156-f053-48ee-abef-6c94e28c392b req-ef8aa07e-b402-4f7e-9444-aace1ddb83b8 service nova] Acquiring lock "5e5f43e8-ce45-45c7-a244-bfa724ea5e3c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.686908] env[62385]: DEBUG oslo_concurrency.lockutils [req-ffd3d156-f053-48ee-abef-6c94e28c392b req-ef8aa07e-b402-4f7e-9444-aace1ddb83b8 service nova] Lock "5e5f43e8-ce45-45c7-a244-bfa724ea5e3c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.687035] env[62385]: DEBUG oslo_concurrency.lockutils [req-ffd3d156-f053-48ee-abef-6c94e28c392b req-ef8aa07e-b402-4f7e-9444-aace1ddb83b8 service nova] Lock "5e5f43e8-ce45-45c7-a244-bfa724ea5e3c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.687275] env[62385]: DEBUG nova.compute.manager [req-ffd3d156-f053-48ee-abef-6c94e28c392b req-ef8aa07e-b402-4f7e-9444-aace1ddb83b8 service nova] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] No waiting events found dispatching network-vif-plugged-823867e4-d6b8-42d1-8597-3c27f6190ab1 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 850.687423] env[62385]: WARNING nova.compute.manager [req-ffd3d156-f053-48ee-abef-6c94e28c392b req-ef8aa07e-b402-4f7e-9444-aace1ddb83b8 service nova] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Received unexpected event network-vif-plugged-823867e4-d6b8-42d1-8597-3c27f6190ab1 for instance with vm_state building and task_state spawning. [ 850.687589] env[62385]: DEBUG nova.compute.manager [req-ffd3d156-f053-48ee-abef-6c94e28c392b req-ef8aa07e-b402-4f7e-9444-aace1ddb83b8 service nova] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Received event network-changed-823867e4-d6b8-42d1-8597-3c27f6190ab1 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 850.687739] env[62385]: DEBUG nova.compute.manager [req-ffd3d156-f053-48ee-abef-6c94e28c392b req-ef8aa07e-b402-4f7e-9444-aace1ddb83b8 service nova] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Refreshing instance network info cache due to event network-changed-823867e4-d6b8-42d1-8597-3c27f6190ab1. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 850.687900] env[62385]: DEBUG oslo_concurrency.lockutils [req-ffd3d156-f053-48ee-abef-6c94e28c392b req-ef8aa07e-b402-4f7e-9444-aace1ddb83b8 service nova] Acquiring lock "refresh_cache-5e5f43e8-ce45-45c7-a244-bfa724ea5e3c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.689822] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-f2cd818c-6b68-4f26-b95e-028548504a6d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 850.690060] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-f2cd818c-6b68-4f26-b95e-028548504a6d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Deleting contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 850.690279] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2cd818c-6b68-4f26-b95e-028548504a6d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Deleting the datastore file [datastore2] 5a82fc98-e656-4476-b7dc-99466e9d6afd {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 850.690565] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-391e7d60-766b-49a7-81fa-09926e15a488 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.703000] env[62385]: DEBUG oslo_vmware.api [None req-f2cd818c-6b68-4f26-b95e-028548504a6d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 850.703000] env[62385]: value = "task-1205889" [ 850.703000] env[62385]: _type = "Task" [ 850.703000] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.723237] env[62385]: DEBUG oslo_vmware.api [None req-f2cd818c-6b68-4f26-b95e-028548504a6d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205889, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.729443] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.737370] env[62385]: DEBUG oslo_vmware.api [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1205887, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.746330] env[62385]: DEBUG oslo_vmware.api [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205885, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.647875} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.746330] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 2a579d13-5372-4340-b7b3-cc02c1912624/2a579d13-5372-4340-b7b3-cc02c1912624.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 850.746866] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 850.746986] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6431fd39-4c2b-4aae-8f08-a5c55896c5a7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.752351] env[62385]: DEBUG nova.network.neutron [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Successfully created port: 73603eea-2766-4ac4-a133-015773484ffc {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 850.756367] env[62385]: DEBUG oslo_vmware.api [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for the task: (returnval){ [ 850.756367] env[62385]: value = "task-1205890" [ 850.756367] env[62385]: _type = "Task" [ 850.756367] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.763922] env[62385]: DEBUG oslo_vmware.api [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205890, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.801776] env[62385]: DEBUG nova.compute.manager [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 850.925131] env[62385]: DEBUG nova.scheduler.client.report [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 850.956935] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "a58cb15f-e4df-4b1e-b09c-2f61dce7200f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.957221] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "a58cb15f-e4df-4b1e-b09c-2f61dce7200f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.015497] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Releasing lock "refresh_cache-5e5f43e8-ce45-45c7-a244-bfa724ea5e3c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.016139] env[62385]: DEBUG nova.compute.manager [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Instance network_info: |[{"id": "823867e4-d6b8-42d1-8597-3c27f6190ab1", "address": "fa:16:3e:56:b4:61", "network": {"id": "b563b70c-d3f1-4f82-97e9-defaf3f9d14b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-55206131-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "8eaa81326f084441a87ac05796958abb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa", "external-id": "nsx-vlan-transportzone-132", "segmentation_id": 132, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap823867e4-d6", "ovs_interfaceid": "823867e4-d6b8-42d1-8597-3c27f6190ab1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 851.016273] env[62385]: DEBUG oslo_concurrency.lockutils [req-ffd3d156-f053-48ee-abef-6c94e28c392b req-ef8aa07e-b402-4f7e-9444-aace1ddb83b8 service nova] Acquired lock "refresh_cache-5e5f43e8-ce45-45c7-a244-bfa724ea5e3c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.016467] env[62385]: DEBUG nova.network.neutron [req-ffd3d156-f053-48ee-abef-6c94e28c392b req-ef8aa07e-b402-4f7e-9444-aace1ddb83b8 service nova] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Refreshing network info cache for port 823867e4-d6b8-42d1-8597-3c27f6190ab1 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 851.018254] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:56:b4:61', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '823867e4-d6b8-42d1-8597-3c27f6190ab1', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 851.028567] env[62385]: DEBUG oslo.service.loopingcall [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 851.029831] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 851.030142] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6162e27f-5249-440b-a583-d9aa03e9eec1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.057069] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 851.057069] env[62385]: value = "task-1205891" [ 851.057069] env[62385]: _type = "Task" [ 851.057069] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.069419] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205891, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.220026] env[62385]: DEBUG oslo_vmware.api [None req-f2cd818c-6b68-4f26-b95e-028548504a6d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205889, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.247451} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.220380] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2cd818c-6b68-4f26-b95e-028548504a6d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 851.220640] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-f2cd818c-6b68-4f26-b95e-028548504a6d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Deleted contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 851.220881] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-f2cd818c-6b68-4f26-b95e-028548504a6d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 851.221142] env[62385]: INFO nova.compute.manager [None req-f2cd818c-6b68-4f26-b95e-028548504a6d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Took 1.17 seconds to destroy the instance on the hypervisor. [ 851.221470] env[62385]: DEBUG oslo.service.loopingcall [None req-f2cd818c-6b68-4f26-b95e-028548504a6d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 851.221758] env[62385]: DEBUG nova.compute.manager [-] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 851.221875] env[62385]: DEBUG nova.network.neutron [-] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 851.234547] env[62385]: DEBUG oslo_vmware.api [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1205887, 'name': PowerOnVM_Task, 'duration_secs': 0.726579} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.234853] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 851.235093] env[62385]: INFO nova.compute.manager [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Took 9.30 seconds to spawn the instance on the hypervisor. [ 851.235300] env[62385]: DEBUG nova.compute.manager [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 851.236272] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9982010-b981-42b0-91bb-6f161d0fe391 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.266088] env[62385]: DEBUG oslo_vmware.api [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205890, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.247071} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.266398] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 851.267423] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e48a2ab-2fde-4da1-81e8-31d162e42116 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.290946] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] 2a579d13-5372-4340-b7b3-cc02c1912624/2a579d13-5372-4340-b7b3-cc02c1912624.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 851.293658] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1856df30-b52e-44ef-859c-cdbcbd364bf3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.319149] env[62385]: DEBUG oslo_vmware.api [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for the task: (returnval){ [ 851.319149] env[62385]: value = "task-1205892" [ 851.319149] env[62385]: _type = "Task" [ 851.319149] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.331651] env[62385]: DEBUG oslo_vmware.api [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205892, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.431425] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.665s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.432012] env[62385]: DEBUG nova.compute.manager [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 851.434850] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.142s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.436731] env[62385]: INFO nova.compute.claims [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 851.461149] env[62385]: DEBUG nova.compute.manager [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 851.538513] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Acquiring lock "e9bf98d1-089a-4078-b2f8-eb77943731e5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.538921] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Lock "e9bf98d1-089a-4078-b2f8-eb77943731e5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.567545] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205891, 'name': CreateVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.583798] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f6418c2e-5dfb-48b5-a327-6d9761dcf8a1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Acquiring lock "d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.584247] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f6418c2e-5dfb-48b5-a327-6d9761dcf8a1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.584491] env[62385]: DEBUG nova.compute.manager [None req-f6418c2e-5dfb-48b5-a327-6d9761dcf8a1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 851.585538] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73675da2-8f4a-4f50-aed6-ee1677420fd5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.593533] env[62385]: DEBUG nova.compute.manager [None req-f6418c2e-5dfb-48b5-a327-6d9761dcf8a1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62385) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 851.595216] env[62385]: DEBUG nova.objects.instance [None req-f6418c2e-5dfb-48b5-a327-6d9761dcf8a1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lazy-loading 'flavor' on Instance uuid d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 851.758562] env[62385]: INFO nova.compute.manager [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Took 50.19 seconds to build instance. [ 851.820047] env[62385]: DEBUG nova.compute.manager [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 851.838360] env[62385]: DEBUG oslo_vmware.api [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205892, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.867263] env[62385]: DEBUG nova.virt.hardware [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 851.871040] env[62385]: DEBUG nova.virt.hardware [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 851.871040] env[62385]: DEBUG nova.virt.hardware [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 851.871040] env[62385]: DEBUG nova.virt.hardware [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 851.871040] env[62385]: DEBUG nova.virt.hardware [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 851.871040] env[62385]: DEBUG nova.virt.hardware [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 851.871040] env[62385]: DEBUG nova.virt.hardware [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 851.871040] env[62385]: DEBUG nova.virt.hardware [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 851.871040] env[62385]: DEBUG nova.virt.hardware [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 851.871040] env[62385]: DEBUG nova.virt.hardware [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 851.871040] env[62385]: DEBUG nova.virt.hardware [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 851.872811] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92e539b5-11e9-41d6-9e77-c52292a49855 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.876543] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d33d6bc6-7bf9-4779-9990-fdee06785b9d tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Acquiring lock "7f2fe830-e16d-4684-91a4-aab219468e77" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.877069] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d33d6bc6-7bf9-4779-9990-fdee06785b9d tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Lock "7f2fe830-e16d-4684-91a4-aab219468e77" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.877462] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d33d6bc6-7bf9-4779-9990-fdee06785b9d tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Acquiring lock "7f2fe830-e16d-4684-91a4-aab219468e77-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.877884] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d33d6bc6-7bf9-4779-9990-fdee06785b9d tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Lock "7f2fe830-e16d-4684-91a4-aab219468e77-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.878292] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d33d6bc6-7bf9-4779-9990-fdee06785b9d tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Lock "7f2fe830-e16d-4684-91a4-aab219468e77-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.881087] env[62385]: INFO nova.compute.manager [None req-d33d6bc6-7bf9-4779-9990-fdee06785b9d tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Terminating instance [ 851.887155] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d33d6bc6-7bf9-4779-9990-fdee06785b9d tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Acquiring lock "refresh_cache-7f2fe830-e16d-4684-91a4-aab219468e77" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.887155] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d33d6bc6-7bf9-4779-9990-fdee06785b9d tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Acquired lock "refresh_cache-7f2fe830-e16d-4684-91a4-aab219468e77" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.887155] env[62385]: DEBUG nova.network.neutron [None req-d33d6bc6-7bf9-4779-9990-fdee06785b9d tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 851.888927] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e0dadc1-3c91-4ff0-bccf-57a6b7ae089b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.941029] env[62385]: DEBUG nova.compute.utils [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 851.945808] env[62385]: DEBUG nova.compute.manager [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 851.945808] env[62385]: DEBUG nova.network.neutron [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 851.981277] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.024205] env[62385]: DEBUG nova.compute.manager [req-d4571d83-34c6-45d6-9778-194a15181878 req-32917306-2662-48cc-ba52-5f199388c1ac service nova] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Received event network-vif-deleted-844976e8-2627-499f-9b01-284d376c86fa {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 852.024369] env[62385]: INFO nova.compute.manager [req-d4571d83-34c6-45d6-9778-194a15181878 req-32917306-2662-48cc-ba52-5f199388c1ac service nova] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Neutron deleted interface 844976e8-2627-499f-9b01-284d376c86fa; detaching it from the instance and deleting it from the info cache [ 852.025029] env[62385]: DEBUG nova.network.neutron [req-d4571d83-34c6-45d6-9778-194a15181878 req-32917306-2662-48cc-ba52-5f199388c1ac service nova] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.037369] env[62385]: DEBUG nova.network.neutron [req-ffd3d156-f053-48ee-abef-6c94e28c392b req-ef8aa07e-b402-4f7e-9444-aace1ddb83b8 service nova] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Updated VIF entry in instance network info cache for port 823867e4-d6b8-42d1-8597-3c27f6190ab1. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 852.037721] env[62385]: DEBUG nova.network.neutron [req-ffd3d156-f053-48ee-abef-6c94e28c392b req-ef8aa07e-b402-4f7e-9444-aace1ddb83b8 service nova] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Updating instance_info_cache with network_info: [{"id": "823867e4-d6b8-42d1-8597-3c27f6190ab1", "address": "fa:16:3e:56:b4:61", "network": {"id": "b563b70c-d3f1-4f82-97e9-defaf3f9d14b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-55206131-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "8eaa81326f084441a87ac05796958abb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa", "external-id": "nsx-vlan-transportzone-132", "segmentation_id": 132, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap823867e4-d6", "ovs_interfaceid": "823867e4-d6b8-42d1-8597-3c27f6190ab1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.049246] env[62385]: DEBUG nova.policy [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '825ffbadf064473aa5eda331914306f2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7ac0c8638f064daba042ded4de8c1cc5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 852.070420] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205891, 'name': CreateVM_Task, 'duration_secs': 0.577834} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.070617] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 852.071406] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.071574] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.071950] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 852.072530] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-37fa8bf6-9b7a-4a54-a4ae-1a0d4df85987 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.077719] env[62385]: DEBUG oslo_vmware.api [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for the task: (returnval){ [ 852.077719] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]523523b5-5d5f-7c46-a19d-69657b570b0d" [ 852.077719] env[62385]: _type = "Task" [ 852.077719] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.086201] env[62385]: DEBUG oslo_vmware.api [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]523523b5-5d5f-7c46-a19d-69657b570b0d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.103247] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6418c2e-5dfb-48b5-a327-6d9761dcf8a1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 852.103501] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a8510d24-b593-47c5-b6f7-2979f37b9bbb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.116331] env[62385]: DEBUG oslo_vmware.api [None req-f6418c2e-5dfb-48b5-a327-6d9761dcf8a1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for the task: (returnval){ [ 852.116331] env[62385]: value = "task-1205893" [ 852.116331] env[62385]: _type = "Task" [ 852.116331] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.127313] env[62385]: DEBUG oslo_vmware.api [None req-f6418c2e-5dfb-48b5-a327-6d9761dcf8a1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1205893, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.261213] env[62385]: DEBUG oslo_concurrency.lockutils [None req-472fb83a-2d98-41c1-a8d3-0a8edb778c36 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "b73a31df-53c8-4550-bf75-5cf3b5aff86c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 57.411s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.323608] env[62385]: DEBUG nova.network.neutron [-] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.339364] env[62385]: DEBUG oslo_vmware.api [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205892, 'name': ReconfigVM_Task, 'duration_secs': 0.604772} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.340997] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Reconfigured VM instance instance-00000049 to attach disk [datastore1] 2a579d13-5372-4340-b7b3-cc02c1912624/2a579d13-5372-4340-b7b3-cc02c1912624.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 852.340997] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c3e23448-a108-4d13-86c6-fdf268cacd59 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.349423] env[62385]: DEBUG oslo_vmware.api [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for the task: (returnval){ [ 852.349423] env[62385]: value = "task-1205894" [ 852.349423] env[62385]: _type = "Task" [ 852.349423] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.366801] env[62385]: DEBUG oslo_vmware.api [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205894, 'name': Rename_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.379065] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Acquiring lock "4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.379351] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Lock "4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.439979] env[62385]: DEBUG nova.network.neutron [None req-d33d6bc6-7bf9-4779-9990-fdee06785b9d tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 852.446008] env[62385]: DEBUG nova.compute.manager [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 852.522183] env[62385]: DEBUG nova.network.neutron [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Successfully updated port: 73603eea-2766-4ac4-a133-015773484ffc {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 852.529739] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d3a7d177-758f-462d-b94d-cd2790da0e98 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.540150] env[62385]: DEBUG oslo_concurrency.lockutils [req-ffd3d156-f053-48ee-abef-6c94e28c392b req-ef8aa07e-b402-4f7e-9444-aace1ddb83b8 service nova] Releasing lock "refresh_cache-5e5f43e8-ce45-45c7-a244-bfa724ea5e3c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.544198] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c96f577-7107-4296-995b-595764ca783b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.579131] env[62385]: DEBUG nova.compute.manager [req-d4571d83-34c6-45d6-9778-194a15181878 req-32917306-2662-48cc-ba52-5f199388c1ac service nova] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Detach interface failed, port_id=844976e8-2627-499f-9b01-284d376c86fa, reason: Instance 5a82fc98-e656-4476-b7dc-99466e9d6afd could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 852.591782] env[62385]: DEBUG oslo_vmware.api [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]523523b5-5d5f-7c46-a19d-69657b570b0d, 'name': SearchDatastore_Task, 'duration_secs': 0.015863} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.592112] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.592356] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 852.592588] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.593171] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.593171] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 852.593300] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-78049cef-9932-4e76-85a7-90c9c3683f59 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.604401] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 852.604555] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 852.605331] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-18f61e4a-8242-486d-a929-e8806ec01fc6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.613678] env[62385]: DEBUG oslo_vmware.api [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for the task: (returnval){ [ 852.613678] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]527614f2-bd8a-c885-7185-728a483e414d" [ 852.613678] env[62385]: _type = "Task" [ 852.613678] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.618093] env[62385]: DEBUG nova.network.neutron [None req-d33d6bc6-7bf9-4779-9990-fdee06785b9d tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.632081] env[62385]: DEBUG oslo_vmware.api [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]527614f2-bd8a-c885-7185-728a483e414d, 'name': SearchDatastore_Task, 'duration_secs': 0.012933} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.632327] env[62385]: DEBUG oslo_vmware.api [None req-f6418c2e-5dfb-48b5-a327-6d9761dcf8a1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1205893, 'name': PowerOffVM_Task, 'duration_secs': 0.290282} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.633746] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6418c2e-5dfb-48b5-a327-6d9761dcf8a1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 852.633928] env[62385]: DEBUG nova.compute.manager [None req-f6418c2e-5dfb-48b5-a327-6d9761dcf8a1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 852.634599] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c3703ab2-5c48-4289-bac9-e801c933bcb6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.637446] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67d50800-ebdc-4d2a-9982-cd853786c8c7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.651864] env[62385]: DEBUG oslo_vmware.api [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for the task: (returnval){ [ 852.651864] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52b03f78-4059-e820-b8c6-e52e2918733b" [ 852.651864] env[62385]: _type = "Task" [ 852.651864] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.666248] env[62385]: DEBUG oslo_vmware.api [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52b03f78-4059-e820-b8c6-e52e2918733b, 'name': SearchDatastore_Task, 'duration_secs': 0.012641} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.666248] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.666248] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c/5e5f43e8-ce45-45c7-a244-bfa724ea5e3c.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 852.666248] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b78d95b1-19ea-4346-ac80-1575e389e7a7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.672287] env[62385]: DEBUG oslo_vmware.api [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for the task: (returnval){ [ 852.672287] env[62385]: value = "task-1205895" [ 852.672287] env[62385]: _type = "Task" [ 852.672287] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.482286] env[62385]: DEBUG nova.network.neutron [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Successfully created port: 2e49e027-be2a-4225-9a34-1e2f9cfb6022 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 853.486639] env[62385]: DEBUG nova.compute.manager [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 853.493027] env[62385]: INFO nova.compute.manager [-] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Took 2.27 seconds to deallocate network for instance. [ 853.493027] env[62385]: DEBUG oslo_concurrency.lockutils [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Acquiring lock "refresh_cache-699a57a9-5a1a-4cd1-8449-723400d9caf4" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.493027] env[62385]: DEBUG oslo_concurrency.lockutils [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Acquired lock "refresh_cache-699a57a9-5a1a-4cd1-8449-723400d9caf4" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.493027] env[62385]: DEBUG nova.network.neutron [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 853.496207] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d33d6bc6-7bf9-4779-9990-fdee06785b9d tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Releasing lock "refresh_cache-7f2fe830-e16d-4684-91a4-aab219468e77" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.496207] env[62385]: DEBUG nova.compute.manager [None req-d33d6bc6-7bf9-4779-9990-fdee06785b9d tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 853.496207] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d33d6bc6-7bf9-4779-9990-fdee06785b9d tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 853.496472] env[62385]: DEBUG nova.compute.manager [req-0094f420-1943-4430-b1ca-0937a5864b6b req-0d8623ab-ac39-4dce-99f3-7211519c454d service nova] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Received event network-vif-plugged-73603eea-2766-4ac4-a133-015773484ffc {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 853.496642] env[62385]: DEBUG oslo_concurrency.lockutils [req-0094f420-1943-4430-b1ca-0937a5864b6b req-0d8623ab-ac39-4dce-99f3-7211519c454d service nova] Acquiring lock "699a57a9-5a1a-4cd1-8449-723400d9caf4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.496866] env[62385]: DEBUG oslo_concurrency.lockutils [req-0094f420-1943-4430-b1ca-0937a5864b6b req-0d8623ab-ac39-4dce-99f3-7211519c454d service nova] Lock "699a57a9-5a1a-4cd1-8449-723400d9caf4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.497074] env[62385]: DEBUG oslo_concurrency.lockutils [req-0094f420-1943-4430-b1ca-0937a5864b6b req-0d8623ab-ac39-4dce-99f3-7211519c454d service nova] Lock "699a57a9-5a1a-4cd1-8449-723400d9caf4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.497310] env[62385]: DEBUG nova.compute.manager [req-0094f420-1943-4430-b1ca-0937a5864b6b req-0d8623ab-ac39-4dce-99f3-7211519c454d service nova] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] No waiting events found dispatching network-vif-plugged-73603eea-2766-4ac4-a133-015773484ffc {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 853.497487] env[62385]: WARNING nova.compute.manager [req-0094f420-1943-4430-b1ca-0937a5864b6b req-0d8623ab-ac39-4dce-99f3-7211519c454d service nova] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Received unexpected event network-vif-plugged-73603eea-2766-4ac4-a133-015773484ffc for instance with vm_state building and task_state spawning. [ 853.497644] env[62385]: DEBUG nova.compute.manager [req-0094f420-1943-4430-b1ca-0937a5864b6b req-0d8623ab-ac39-4dce-99f3-7211519c454d service nova] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Received event network-changed-73603eea-2766-4ac4-a133-015773484ffc {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 853.497803] env[62385]: DEBUG nova.compute.manager [req-0094f420-1943-4430-b1ca-0937a5864b6b req-0d8623ab-ac39-4dce-99f3-7211519c454d service nova] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Refreshing instance network info cache due to event network-changed-73603eea-2766-4ac4-a133-015773484ffc. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 853.497963] env[62385]: DEBUG oslo_concurrency.lockutils [req-0094f420-1943-4430-b1ca-0937a5864b6b req-0d8623ab-ac39-4dce-99f3-7211519c454d service nova] Acquiring lock "refresh_cache-699a57a9-5a1a-4cd1-8449-723400d9caf4" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.500859] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f6418c2e-5dfb-48b5-a327-6d9761dcf8a1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.917s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.507218] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6288396-d639-4d93-a40b-6ac741bec82e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.521182] env[62385]: DEBUG oslo_vmware.api [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205895, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.676812} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.521850] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c/5e5f43e8-ce45-45c7-a244-bfa724ea5e3c.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 853.522353] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 853.522420] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-497ef2c9-7f79-499a-951a-7fda45fe0652 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.530894] env[62385]: DEBUG oslo_vmware.api [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205894, 'name': Rename_Task, 'duration_secs': 0.203721} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.531545] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d33d6bc6-7bf9-4779-9990-fdee06785b9d tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 853.535145] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 853.536202] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f7c77b7d-11ac-4252-8116-700771859797 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.541054] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-df77ee23-72d0-4ad1-9c7b-9fc763e76aca {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.542479] env[62385]: DEBUG oslo_vmware.api [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for the task: (returnval){ [ 853.542479] env[62385]: value = "task-1205896" [ 853.542479] env[62385]: _type = "Task" [ 853.542479] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.550663] env[62385]: DEBUG oslo_vmware.api [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for the task: (returnval){ [ 853.550663] env[62385]: value = "task-1205898" [ 853.550663] env[62385]: _type = "Task" [ 853.550663] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.550993] env[62385]: DEBUG oslo_vmware.api [None req-d33d6bc6-7bf9-4779-9990-fdee06785b9d tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Waiting for the task: (returnval){ [ 853.550993] env[62385]: value = "task-1205897" [ 853.550993] env[62385]: _type = "Task" [ 853.550993] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.569320] env[62385]: DEBUG oslo_vmware.api [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205896, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.575750] env[62385]: DEBUG oslo_vmware.api [None req-d33d6bc6-7bf9-4779-9990-fdee06785b9d tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': task-1205897, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.579114] env[62385]: DEBUG oslo_vmware.api [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205898, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.731940] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fb07241-35ba-4574-b51f-3ae6bb9528ff {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.740680] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff2a336f-b06a-4407-af86-d383d46fc21e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.773190] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3df68ee5-0dc2-4bd3-aeba-d2ce7dd00629 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.781364] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b34e99fc-4c31-498d-a4e8-96c0915098bd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.794920] env[62385]: DEBUG nova.compute.provider_tree [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 853.820166] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Acquiring lock "2bcd9457-8a73-4e7a-b778-d52c468b3aae" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.820234] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Lock "2bcd9457-8a73-4e7a-b778-d52c468b3aae" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.009349] env[62385]: DEBUG nova.compute.manager [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 854.022142] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f2cd818c-6b68-4f26-b95e-028548504a6d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.029693] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.041452] env[62385]: DEBUG nova.virt.hardware [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 854.041452] env[62385]: DEBUG nova.virt.hardware [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 854.041452] env[62385]: DEBUG nova.virt.hardware [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 854.041452] env[62385]: DEBUG nova.virt.hardware [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 854.041452] env[62385]: DEBUG nova.virt.hardware [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 854.041452] env[62385]: DEBUG nova.virt.hardware [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 854.041452] env[62385]: DEBUG nova.virt.hardware [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 854.041452] env[62385]: DEBUG nova.virt.hardware [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 854.041907] env[62385]: DEBUG nova.virt.hardware [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 854.041907] env[62385]: DEBUG nova.virt.hardware [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 854.041971] env[62385]: DEBUG nova.virt.hardware [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 854.042815] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efe502ab-b256-4535-871c-c4bbd24843cf {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.052177] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2bb6b7b6-ebba-481c-ae8f-c30766e091ab tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "b73a31df-53c8-4550-bf75-5cf3b5aff86c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.052177] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2bb6b7b6-ebba-481c-ae8f-c30766e091ab tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "b73a31df-53c8-4550-bf75-5cf3b5aff86c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.052177] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2bb6b7b6-ebba-481c-ae8f-c30766e091ab tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "b73a31df-53c8-4550-bf75-5cf3b5aff86c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.052177] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2bb6b7b6-ebba-481c-ae8f-c30766e091ab tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "b73a31df-53c8-4550-bf75-5cf3b5aff86c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.052177] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2bb6b7b6-ebba-481c-ae8f-c30766e091ab tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "b73a31df-53c8-4550-bf75-5cf3b5aff86c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.060723] env[62385]: DEBUG oslo_vmware.api [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205896, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076798} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.061357] env[62385]: INFO nova.compute.manager [None req-2bb6b7b6-ebba-481c-ae8f-c30766e091ab tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Terminating instance [ 854.070266] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98650a3b-b9cc-4f3d-99c0-832e6becb262 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.074994] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 854.076268] env[62385]: DEBUG nova.network.neutron [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 854.079940] env[62385]: DEBUG nova.compute.manager [None req-2bb6b7b6-ebba-481c-ae8f-c30766e091ab tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 854.079940] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-2bb6b7b6-ebba-481c-ae8f-c30766e091ab tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 854.079940] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4805642-5b22-4054-b9e9-9ec5ec72494a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.083018] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5522017-917f-4f49-a312-0b38fc51a3c8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.096900] env[62385]: DEBUG oslo_vmware.api [None req-d33d6bc6-7bf9-4779-9990-fdee06785b9d tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': task-1205897, 'name': PowerOffVM_Task, 'duration_secs': 0.166417} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.097209] env[62385]: DEBUG oslo_vmware.api [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205898, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.128740] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d33d6bc6-7bf9-4779-9990-fdee06785b9d tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 854.129067] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d33d6bc6-7bf9-4779-9990-fdee06785b9d tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 854.139629] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c/5e5f43e8-ce45-45c7-a244-bfa724ea5e3c.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 854.143028] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bd17253d-8bd1-41c3-9700-bd3db3e0d61a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.144815] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-993509cb-8dd4-45a0-b9bd-67f0a1d9f90b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.166210] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bb6b7b6-ebba-481c-ae8f-c30766e091ab tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 854.167405] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b9636a72-45e1-463e-a50f-53b806977a90 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.172696] env[62385]: DEBUG oslo_vmware.api [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for the task: (returnval){ [ 854.172696] env[62385]: value = "task-1205900" [ 854.172696] env[62385]: _type = "Task" [ 854.172696] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.177740] env[62385]: DEBUG oslo_vmware.api [None req-2bb6b7b6-ebba-481c-ae8f-c30766e091ab tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 854.177740] env[62385]: value = "task-1205901" [ 854.177740] env[62385]: _type = "Task" [ 854.177740] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.185826] env[62385]: DEBUG oslo_vmware.api [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205900, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.193804] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d33d6bc6-7bf9-4779-9990-fdee06785b9d tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 854.194155] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d33d6bc6-7bf9-4779-9990-fdee06785b9d tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 854.194392] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-d33d6bc6-7bf9-4779-9990-fdee06785b9d tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Deleting the datastore file [datastore1] 7f2fe830-e16d-4684-91a4-aab219468e77 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 854.194743] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5667060f-e818-461d-b3cb-53736e4118c9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.199455] env[62385]: DEBUG oslo_vmware.api [None req-2bb6b7b6-ebba-481c-ae8f-c30766e091ab tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1205901, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.203826] env[62385]: DEBUG oslo_vmware.api [None req-d33d6bc6-7bf9-4779-9990-fdee06785b9d tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Waiting for the task: (returnval){ [ 854.203826] env[62385]: value = "task-1205902" [ 854.203826] env[62385]: _type = "Task" [ 854.203826] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.213559] env[62385]: DEBUG oslo_vmware.api [None req-d33d6bc6-7bf9-4779-9990-fdee06785b9d tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': task-1205902, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.278694] env[62385]: DEBUG oslo_vmware.rw_handles [None req-a6e09c15-369b-4d28-968e-f207336024dc tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52926248-cc62-db81-3e11-08a35f3299a4/disk-0.vmdk. {{(pid=62385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 854.278694] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87544432-5c1d-4fed-ac0e-54f432fe112b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.286423] env[62385]: DEBUG oslo_vmware.rw_handles [None req-a6e09c15-369b-4d28-968e-f207336024dc tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52926248-cc62-db81-3e11-08a35f3299a4/disk-0.vmdk is in state: ready. {{(pid=62385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 854.286548] env[62385]: ERROR oslo_vmware.rw_handles [None req-a6e09c15-369b-4d28-968e-f207336024dc tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52926248-cc62-db81-3e11-08a35f3299a4/disk-0.vmdk due to incomplete transfer. [ 854.286755] env[62385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-070d2cda-8669-4348-9681-73e957ef4b3c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.294164] env[62385]: DEBUG oslo_vmware.rw_handles [None req-a6e09c15-369b-4d28-968e-f207336024dc tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52926248-cc62-db81-3e11-08a35f3299a4/disk-0.vmdk. {{(pid=62385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 854.294395] env[62385]: DEBUG nova.virt.vmwareapi.images [None req-a6e09c15-369b-4d28-968e-f207336024dc tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Uploaded image 496e7b4f-f413-4c90-8a25-d4ab00a67dbf to the Glance image server {{(pid=62385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 854.297036] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6e09c15-369b-4d28-968e-f207336024dc tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Destroying the VM {{(pid=62385) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 854.297816] env[62385]: DEBUG nova.scheduler.client.report [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 854.301083] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-aa779242-1cfe-42a9-b158-814f0a69b3db {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.308182] env[62385]: DEBUG oslo_vmware.api [None req-a6e09c15-369b-4d28-968e-f207336024dc tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Waiting for the task: (returnval){ [ 854.308182] env[62385]: value = "task-1205903" [ 854.308182] env[62385]: _type = "Task" [ 854.308182] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.317661] env[62385]: DEBUG oslo_vmware.api [None req-a6e09c15-369b-4d28-968e-f207336024dc tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205903, 'name': Destroy_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.405029] env[62385]: DEBUG nova.objects.instance [None req-1d9c956f-96c7-4e9f-8bf2-e0890adf91a5 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lazy-loading 'flavor' on Instance uuid d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 854.409160] env[62385]: DEBUG nova.network.neutron [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Updating instance_info_cache with network_info: [{"id": "73603eea-2766-4ac4-a133-015773484ffc", "address": "fa:16:3e:94:e9:8a", "network": {"id": "78235fdc-e80c-4df9-b1fa-c4dac60e3db4", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-51998007-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c556f4359bd44d2f868a796249b25694", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1cb94a1a-f287-46e7-b63b-ec692c2141b4", "external-id": "nsx-vlan-transportzone-346", "segmentation_id": 346, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73603eea-27", "ovs_interfaceid": "73603eea-2766-4ac4-a133-015773484ffc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.572171] env[62385]: DEBUG oslo_vmware.api [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205898, 'name': PowerOnVM_Task, 'duration_secs': 0.558226} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.576018] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 854.576018] env[62385]: INFO nova.compute.manager [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Took 10.09 seconds to spawn the instance on the hypervisor. [ 854.576018] env[62385]: DEBUG nova.compute.manager [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 854.576018] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-330aa753-2e68-4d9d-b3e4-a7f56e2c819b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.684998] env[62385]: DEBUG oslo_vmware.api [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205900, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.690146] env[62385]: DEBUG oslo_vmware.api [None req-2bb6b7b6-ebba-481c-ae8f-c30766e091ab tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1205901, 'name': PowerOffVM_Task, 'duration_secs': 0.401688} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.690388] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bb6b7b6-ebba-481c-ae8f-c30766e091ab tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 854.690562] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-2bb6b7b6-ebba-481c-ae8f-c30766e091ab tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 854.690804] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fdc31b57-1289-4d4e-8c1f-b0a21eefa5ea {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.713334] env[62385]: DEBUG oslo_vmware.api [None req-d33d6bc6-7bf9-4779-9990-fdee06785b9d tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Task: {'id': task-1205902, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.314774} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.713577] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-d33d6bc6-7bf9-4779-9990-fdee06785b9d tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 854.713761] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d33d6bc6-7bf9-4779-9990-fdee06785b9d tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 854.713945] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d33d6bc6-7bf9-4779-9990-fdee06785b9d tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 854.714137] env[62385]: INFO nova.compute.manager [None req-d33d6bc6-7bf9-4779-9990-fdee06785b9d tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Took 1.22 seconds to destroy the instance on the hypervisor. [ 854.714518] env[62385]: DEBUG oslo.service.loopingcall [None req-d33d6bc6-7bf9-4779-9990-fdee06785b9d tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 854.714731] env[62385]: DEBUG nova.compute.manager [-] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 854.714828] env[62385]: DEBUG nova.network.neutron [-] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 854.736841] env[62385]: DEBUG nova.network.neutron [-] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 854.804511] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.370s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.805191] env[62385]: DEBUG nova.compute.manager [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 854.808305] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.221s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.809765] env[62385]: INFO nova.compute.claims [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 854.818086] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-2bb6b7b6-ebba-481c-ae8f-c30766e091ab tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 854.818312] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-2bb6b7b6-ebba-481c-ae8f-c30766e091ab tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Deleting contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 854.818501] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-2bb6b7b6-ebba-481c-ae8f-c30766e091ab tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Deleting the datastore file [datastore2] b73a31df-53c8-4550-bf75-5cf3b5aff86c {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 854.819200] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-aa6ac3ed-64d5-4ca4-8adf-64bbc2c5ae3b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.824239] env[62385]: DEBUG oslo_vmware.api [None req-a6e09c15-369b-4d28-968e-f207336024dc tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205903, 'name': Destroy_Task, 'duration_secs': 0.368914} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.824766] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-a6e09c15-369b-4d28-968e-f207336024dc tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Destroyed the VM [ 854.825142] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a6e09c15-369b-4d28-968e-f207336024dc tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Deleting Snapshot of the VM instance {{(pid=62385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 854.825323] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-0679ce60-7ef5-4007-ae0f-49c31e4e6f85 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.829793] env[62385]: DEBUG oslo_vmware.api [None req-2bb6b7b6-ebba-481c-ae8f-c30766e091ab tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 854.829793] env[62385]: value = "task-1205905" [ 854.829793] env[62385]: _type = "Task" [ 854.829793] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.834358] env[62385]: DEBUG oslo_vmware.api [None req-a6e09c15-369b-4d28-968e-f207336024dc tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Waiting for the task: (returnval){ [ 854.834358] env[62385]: value = "task-1205906" [ 854.834358] env[62385]: _type = "Task" [ 854.834358] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.837263] env[62385]: DEBUG oslo_vmware.api [None req-2bb6b7b6-ebba-481c-ae8f-c30766e091ab tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1205905, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.845275] env[62385]: DEBUG oslo_vmware.api [None req-a6e09c15-369b-4d28-968e-f207336024dc tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205906, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.909033] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1d9c956f-96c7-4e9f-8bf2-e0890adf91a5 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Acquiring lock "refresh_cache-d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.909126] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1d9c956f-96c7-4e9f-8bf2-e0890adf91a5 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Acquired lock "refresh_cache-d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.910242] env[62385]: DEBUG nova.network.neutron [None req-1d9c956f-96c7-4e9f-8bf2-e0890adf91a5 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 854.910242] env[62385]: DEBUG nova.objects.instance [None req-1d9c956f-96c7-4e9f-8bf2-e0890adf91a5 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lazy-loading 'info_cache' on Instance uuid d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 854.910912] env[62385]: DEBUG oslo_concurrency.lockutils [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Releasing lock "refresh_cache-699a57a9-5a1a-4cd1-8449-723400d9caf4" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.911247] env[62385]: DEBUG nova.compute.manager [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Instance network_info: |[{"id": "73603eea-2766-4ac4-a133-015773484ffc", "address": "fa:16:3e:94:e9:8a", "network": {"id": "78235fdc-e80c-4df9-b1fa-c4dac60e3db4", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-51998007-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c556f4359bd44d2f868a796249b25694", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1cb94a1a-f287-46e7-b63b-ec692c2141b4", "external-id": "nsx-vlan-transportzone-346", "segmentation_id": 346, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73603eea-27", "ovs_interfaceid": "73603eea-2766-4ac4-a133-015773484ffc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 854.911724] env[62385]: DEBUG oslo_concurrency.lockutils [req-0094f420-1943-4430-b1ca-0937a5864b6b req-0d8623ab-ac39-4dce-99f3-7211519c454d service nova] Acquired lock "refresh_cache-699a57a9-5a1a-4cd1-8449-723400d9caf4" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.911938] env[62385]: DEBUG nova.network.neutron [req-0094f420-1943-4430-b1ca-0937a5864b6b req-0d8623ab-ac39-4dce-99f3-7211519c454d service nova] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Refreshing network info cache for port 73603eea-2766-4ac4-a133-015773484ffc {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 854.913388] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:94:e9:8a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1cb94a1a-f287-46e7-b63b-ec692c2141b4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '73603eea-2766-4ac4-a133-015773484ffc', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 854.921532] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Creating folder: Project (c556f4359bd44d2f868a796249b25694). Parent ref: group-v261107. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 854.922145] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-22ff1fbe-95b4-4d1a-9c23-60a356730000 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.933352] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Created folder: Project (c556f4359bd44d2f868a796249b25694) in parent group-v261107. [ 854.933545] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Creating folder: Instances. Parent ref: group-v261223. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 854.933776] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-175a36a9-113b-42d9-85dc-e42181bb74fe {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.943323] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Created folder: Instances in parent group-v261223. [ 854.943518] env[62385]: DEBUG oslo.service.loopingcall [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 854.943727] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 854.943941] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-46566fcb-fa24-4346-a085-52f73eba3c29 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.962787] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 854.962787] env[62385]: value = "task-1205909" [ 854.962787] env[62385]: _type = "Task" [ 854.962787] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.970347] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205909, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.095282] env[62385]: INFO nova.compute.manager [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Took 45.64 seconds to build instance. [ 855.188655] env[62385]: DEBUG oslo_vmware.api [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205900, 'name': ReconfigVM_Task, 'duration_secs': 0.852403} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.189243] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Reconfigured VM instance instance-0000004a to attach disk [datastore1] 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c/5e5f43e8-ce45-45c7-a244-bfa724ea5e3c.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 855.190248] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a462cfdb-09f6-443a-be59-09f3e8a80bad {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.204032] env[62385]: DEBUG oslo_vmware.api [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for the task: (returnval){ [ 855.204032] env[62385]: value = "task-1205910" [ 855.204032] env[62385]: _type = "Task" [ 855.204032] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.220053] env[62385]: DEBUG oslo_vmware.api [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205910, 'name': Rename_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.241095] env[62385]: DEBUG nova.network.neutron [-] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.300770] env[62385]: DEBUG oslo_concurrency.lockutils [None req-073e1e89-6834-4e1e-98dc-44639e7d57b1 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "9b9fd167-d843-4b9d-9e4f-dbde5d3d629e" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.301117] env[62385]: DEBUG oslo_concurrency.lockutils [None req-073e1e89-6834-4e1e-98dc-44639e7d57b1 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "9b9fd167-d843-4b9d-9e4f-dbde5d3d629e" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.301329] env[62385]: DEBUG nova.compute.manager [None req-073e1e89-6834-4e1e-98dc-44639e7d57b1 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 855.302680] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2966aaa-17fd-4c27-81d8-e059f7dbdfa9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.309285] env[62385]: DEBUG nova.compute.manager [None req-073e1e89-6834-4e1e-98dc-44639e7d57b1 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62385) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 855.309954] env[62385]: DEBUG nova.objects.instance [None req-073e1e89-6834-4e1e-98dc-44639e7d57b1 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lazy-loading 'flavor' on Instance uuid 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 855.315493] env[62385]: DEBUG nova.compute.utils [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 855.319732] env[62385]: DEBUG nova.compute.manager [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 855.319921] env[62385]: DEBUG nova.network.neutron [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 855.348468] env[62385]: DEBUG oslo_vmware.api [None req-a6e09c15-369b-4d28-968e-f207336024dc tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205906, 'name': RemoveSnapshot_Task} progress is 43%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.351769] env[62385]: DEBUG oslo_vmware.api [None req-2bb6b7b6-ebba-481c-ae8f-c30766e091ab tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1205905, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.290378} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.352123] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-2bb6b7b6-ebba-481c-ae8f-c30766e091ab tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 855.352219] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-2bb6b7b6-ebba-481c-ae8f-c30766e091ab tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Deleted contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 855.352404] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-2bb6b7b6-ebba-481c-ae8f-c30766e091ab tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 855.352613] env[62385]: INFO nova.compute.manager [None req-2bb6b7b6-ebba-481c-ae8f-c30766e091ab tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Took 1.27 seconds to destroy the instance on the hypervisor. [ 855.352858] env[62385]: DEBUG oslo.service.loopingcall [None req-2bb6b7b6-ebba-481c-ae8f-c30766e091ab tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 855.353089] env[62385]: DEBUG nova.compute.manager [-] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 855.353421] env[62385]: DEBUG nova.network.neutron [-] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 855.414395] env[62385]: DEBUG nova.objects.base [None req-1d9c956f-96c7-4e9f-8bf2-e0890adf91a5 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=62385) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 855.475057] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205909, 'name': CreateVM_Task, 'duration_secs': 0.428781} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.475249] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 855.475902] env[62385]: DEBUG oslo_concurrency.lockutils [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.476083] env[62385]: DEBUG oslo_concurrency.lockutils [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.476425] env[62385]: DEBUG oslo_concurrency.lockutils [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 855.476688] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-08137ac8-691c-4b37-9e4f-d24975c2f87c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.482028] env[62385]: DEBUG oslo_vmware.api [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Waiting for the task: (returnval){ [ 855.482028] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a485a0-c450-df25-78f8-33bf786a2142" [ 855.482028] env[62385]: _type = "Task" [ 855.482028] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.489889] env[62385]: DEBUG oslo_vmware.api [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a485a0-c450-df25-78f8-33bf786a2142, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.598372] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fb8c8815-eee4-42e3-8ae7-8c60063a7e88 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Lock "2a579d13-5372-4340-b7b3-cc02c1912624" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 58.337s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.655840] env[62385]: DEBUG nova.policy [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '923d29b88d064a1da592e1175fd8fe4b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd68661a0197d4e3cacc06ea433ac706e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 855.714726] env[62385]: DEBUG oslo_vmware.api [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205910, 'name': Rename_Task, 'duration_secs': 0.185441} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.714993] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 855.715297] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b4f88e5b-408c-4d79-8185-f0746368d8c3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.722730] env[62385]: DEBUG oslo_vmware.api [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for the task: (returnval){ [ 855.722730] env[62385]: value = "task-1205911" [ 855.722730] env[62385]: _type = "Task" [ 855.722730] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.733320] env[62385]: DEBUG oslo_vmware.api [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205911, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.742821] env[62385]: INFO nova.compute.manager [-] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Took 1.03 seconds to deallocate network for instance. [ 855.798048] env[62385]: DEBUG nova.compute.manager [req-65fc879d-7b4e-47cd-9a64-abd6bde9872a req-69d1d643-1f57-4af2-9cb8-1412f288aa84 service nova] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Received event network-vif-plugged-2e49e027-be2a-4225-9a34-1e2f9cfb6022 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 855.798304] env[62385]: DEBUG oslo_concurrency.lockutils [req-65fc879d-7b4e-47cd-9a64-abd6bde9872a req-69d1d643-1f57-4af2-9cb8-1412f288aa84 service nova] Acquiring lock "b7a286a2-9a19-4ef1-8a46-6e2a04e22413-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.798525] env[62385]: DEBUG oslo_concurrency.lockutils [req-65fc879d-7b4e-47cd-9a64-abd6bde9872a req-69d1d643-1f57-4af2-9cb8-1412f288aa84 service nova] Lock "b7a286a2-9a19-4ef1-8a46-6e2a04e22413-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.798698] env[62385]: DEBUG oslo_concurrency.lockutils [req-65fc879d-7b4e-47cd-9a64-abd6bde9872a req-69d1d643-1f57-4af2-9cb8-1412f288aa84 service nova] Lock "b7a286a2-9a19-4ef1-8a46-6e2a04e22413-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.798901] env[62385]: DEBUG nova.compute.manager [req-65fc879d-7b4e-47cd-9a64-abd6bde9872a req-69d1d643-1f57-4af2-9cb8-1412f288aa84 service nova] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] No waiting events found dispatching network-vif-plugged-2e49e027-be2a-4225-9a34-1e2f9cfb6022 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 855.799472] env[62385]: WARNING nova.compute.manager [req-65fc879d-7b4e-47cd-9a64-abd6bde9872a req-69d1d643-1f57-4af2-9cb8-1412f288aa84 service nova] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Received unexpected event network-vif-plugged-2e49e027-be2a-4225-9a34-1e2f9cfb6022 for instance with vm_state building and task_state spawning. [ 855.814665] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-073e1e89-6834-4e1e-98dc-44639e7d57b1 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 855.814947] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2de200ff-6b87-4d54-94e1-b0de691b5621 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.823596] env[62385]: DEBUG nova.compute.manager [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 855.831021] env[62385]: DEBUG oslo_vmware.api [None req-073e1e89-6834-4e1e-98dc-44639e7d57b1 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 855.831021] env[62385]: value = "task-1205912" [ 855.831021] env[62385]: _type = "Task" [ 855.831021] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.837428] env[62385]: DEBUG oslo_vmware.api [None req-073e1e89-6834-4e1e-98dc-44639e7d57b1 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1205912, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.850150] env[62385]: DEBUG oslo_vmware.api [None req-a6e09c15-369b-4d28-968e-f207336024dc tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205906, 'name': RemoveSnapshot_Task, 'duration_secs': 0.694526} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.853313] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a6e09c15-369b-4d28-968e-f207336024dc tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Deleted Snapshot of the VM instance {{(pid=62385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 855.853976] env[62385]: INFO nova.compute.manager [None req-a6e09c15-369b-4d28-968e-f207336024dc tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Took 15.24 seconds to snapshot the instance on the hypervisor. [ 855.997171] env[62385]: DEBUG oslo_vmware.api [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a485a0-c450-df25-78f8-33bf786a2142, 'name': SearchDatastore_Task, 'duration_secs': 0.035993} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.997171] env[62385]: DEBUG oslo_concurrency.lockutils [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.997171] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 855.997171] env[62385]: DEBUG oslo_concurrency.lockutils [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.997171] env[62385]: DEBUG oslo_concurrency.lockutils [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.997171] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 855.997430] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8f68094f-5818-4e4b-8b3c-021a7de197fc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.002354] env[62385]: DEBUG nova.network.neutron [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Successfully updated port: 2e49e027-be2a-4225-9a34-1e2f9cfb6022 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 856.008380] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 856.008573] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 856.012224] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a8d62b8b-3323-4227-9e15-edbd8c42cf27 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.020098] env[62385]: DEBUG oslo_vmware.api [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Waiting for the task: (returnval){ [ 856.020098] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5293e7af-9eea-2c19-9ad8-9d4f54182484" [ 856.020098] env[62385]: _type = "Task" [ 856.020098] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.035100] env[62385]: DEBUG oslo_vmware.api [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5293e7af-9eea-2c19-9ad8-9d4f54182484, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.103147] env[62385]: DEBUG nova.compute.manager [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 856.124999] env[62385]: DEBUG nova.network.neutron [-] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.209744] env[62385]: DEBUG nova.network.neutron [None req-1d9c956f-96c7-4e9f-8bf2-e0890adf91a5 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Updating instance_info_cache with network_info: [{"id": "2c31d161-814f-4c23-9882-0aa436bb9e03", "address": "fa:16:3e:d7:43:42", "network": {"id": "aca904e1-fbd6-4def-bbec-bd70c0fc8af9", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-249813742-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.141", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e99f1e4cfd84bf083c8261b2ffcfa2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f7abe5d-b8fe-4983-bd50-e7469f1fe7f3", "external-id": "nsx-vlan-transportzone-263", "segmentation_id": 263, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c31d161-81", "ovs_interfaceid": "2c31d161-814f-4c23-9882-0aa436bb9e03", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.235654] env[62385]: DEBUG oslo_vmware.api [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205911, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.250939] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d33d6bc6-7bf9-4779-9990-fdee06785b9d tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.254819] env[62385]: DEBUG nova.network.neutron [req-0094f420-1943-4430-b1ca-0937a5864b6b req-0d8623ab-ac39-4dce-99f3-7211519c454d service nova] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Updated VIF entry in instance network info cache for port 73603eea-2766-4ac4-a133-015773484ffc. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 856.254819] env[62385]: DEBUG nova.network.neutron [req-0094f420-1943-4430-b1ca-0937a5864b6b req-0d8623ab-ac39-4dce-99f3-7211519c454d service nova] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Updating instance_info_cache with network_info: [{"id": "73603eea-2766-4ac4-a133-015773484ffc", "address": "fa:16:3e:94:e9:8a", "network": {"id": "78235fdc-e80c-4df9-b1fa-c4dac60e3db4", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-51998007-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c556f4359bd44d2f868a796249b25694", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1cb94a1a-f287-46e7-b63b-ec692c2141b4", "external-id": "nsx-vlan-transportzone-346", "segmentation_id": 346, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73603eea-27", "ovs_interfaceid": "73603eea-2766-4ac4-a133-015773484ffc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.263953] env[62385]: DEBUG nova.compute.manager [req-522ab897-d963-4e41-a48f-c5bc68b74f9b req-9e4bfa1b-59b3-44f8-b22e-315ece2e8f1c service nova] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Received event network-vif-deleted-508791ff-1aa3-43d6-9dd9-1aac3351b688 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 856.302263] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebef0821-d151-4e20-b065-3172ccff5b1b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.310379] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51a8d14e-999f-4848-b74e-bc2d00de99d3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.349910] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89abf3d1-f06d-44b1-a9c3-0a9fec8d8424 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.361227] env[62385]: DEBUG oslo_vmware.api [None req-073e1e89-6834-4e1e-98dc-44639e7d57b1 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1205912, 'name': PowerOffVM_Task, 'duration_secs': 0.185602} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.363579] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-073e1e89-6834-4e1e-98dc-44639e7d57b1 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 856.363946] env[62385]: DEBUG nova.compute.manager [None req-073e1e89-6834-4e1e-98dc-44639e7d57b1 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 856.364829] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-155d2a2d-e646-49a5-99be-d94ac7d891f2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.369151] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb3a19f7-e58c-42dd-8f1f-33e1c408714f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.387733] env[62385]: DEBUG nova.compute.provider_tree [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 856.486790] env[62385]: DEBUG nova.network.neutron [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Successfully created port: 2572a945-d7b9-47e5-8641-3718434e80ca {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 856.505206] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquiring lock "refresh_cache-b7a286a2-9a19-4ef1-8a46-6e2a04e22413" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.505206] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquired lock "refresh_cache-b7a286a2-9a19-4ef1-8a46-6e2a04e22413" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.505939] env[62385]: DEBUG nova.network.neutron [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 856.531300] env[62385]: DEBUG oslo_vmware.api [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5293e7af-9eea-2c19-9ad8-9d4f54182484, 'name': SearchDatastore_Task, 'duration_secs': 0.032336} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.532143] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0798a02-d8cc-4cd1-8b46-6e5ac4812197 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.537543] env[62385]: DEBUG oslo_vmware.api [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Waiting for the task: (returnval){ [ 856.537543] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52d29b55-eba1-9333-ccb9-44812b49be02" [ 856.537543] env[62385]: _type = "Task" [ 856.537543] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.545804] env[62385]: DEBUG oslo_vmware.api [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52d29b55-eba1-9333-ccb9-44812b49be02, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.627243] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.627690] env[62385]: INFO nova.compute.manager [-] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Took 1.27 seconds to deallocate network for instance. [ 856.649839] env[62385]: DEBUG nova.compute.manager [None req-d86e0c75-5a2a-4394-8269-7f8858af71d8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 856.650813] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c47fd9c6-f91e-40a3-9137-cf983e8472ef {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.712802] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1d9c956f-96c7-4e9f-8bf2-e0890adf91a5 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Releasing lock "refresh_cache-d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.733341] env[62385]: DEBUG oslo_vmware.api [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205911, 'name': PowerOnVM_Task, 'duration_secs': 0.704128} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.733594] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 856.733795] env[62385]: INFO nova.compute.manager [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Took 9.72 seconds to spawn the instance on the hypervisor. [ 856.733979] env[62385]: DEBUG nova.compute.manager [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 856.734845] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81615ad8-277c-4057-952c-8251e5acbbae {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.757356] env[62385]: DEBUG oslo_concurrency.lockutils [req-0094f420-1943-4430-b1ca-0937a5864b6b req-0d8623ab-ac39-4dce-99f3-7211519c454d service nova] Releasing lock "refresh_cache-699a57a9-5a1a-4cd1-8449-723400d9caf4" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.853805] env[62385]: DEBUG nova.compute.manager [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 856.879977] env[62385]: DEBUG nova.virt.hardware [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 856.880288] env[62385]: DEBUG nova.virt.hardware [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 856.880468] env[62385]: DEBUG nova.virt.hardware [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 856.880660] env[62385]: DEBUG nova.virt.hardware [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 856.880813] env[62385]: DEBUG nova.virt.hardware [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 856.880999] env[62385]: DEBUG nova.virt.hardware [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 856.881363] env[62385]: DEBUG nova.virt.hardware [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 856.881546] env[62385]: DEBUG nova.virt.hardware [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 856.881729] env[62385]: DEBUG nova.virt.hardware [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 856.881903] env[62385]: DEBUG nova.virt.hardware [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 856.882095] env[62385]: DEBUG nova.virt.hardware [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 856.882972] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-272032b1-85a5-4200-a8e1-e96c40101830 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.892807] env[62385]: DEBUG nova.scheduler.client.report [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 856.897887] env[62385]: DEBUG oslo_concurrency.lockutils [None req-073e1e89-6834-4e1e-98dc-44639e7d57b1 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "9b9fd167-d843-4b9d-9e4f-dbde5d3d629e" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.597s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.900317] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e95f369-c36b-43da-b8e6-741ca59f9c97 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.944785] env[62385]: DEBUG nova.network.neutron [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Successfully created port: 6820d526-f9c3-43cc-91db-538e9c3bd249 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 857.055652] env[62385]: DEBUG oslo_vmware.api [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52d29b55-eba1-9333-ccb9-44812b49be02, 'name': SearchDatastore_Task, 'duration_secs': 0.019747} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.056986] env[62385]: DEBUG oslo_concurrency.lockutils [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.056986] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 699a57a9-5a1a-4cd1-8449-723400d9caf4/699a57a9-5a1a-4cd1-8449-723400d9caf4.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 857.057687] env[62385]: DEBUG nova.network.neutron [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 857.059850] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-71c53fc3-97bb-4c19-a441-8677a835f8ad {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.071914] env[62385]: DEBUG oslo_vmware.api [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Waiting for the task: (returnval){ [ 857.071914] env[62385]: value = "task-1205913" [ 857.071914] env[62385]: _type = "Task" [ 857.071914] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.083207] env[62385]: DEBUG oslo_vmware.api [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Task: {'id': task-1205913, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.134187] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2bb6b7b6-ebba-481c-ae8f-c30766e091ab tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.163989] env[62385]: INFO nova.compute.manager [None req-d86e0c75-5a2a-4394-8269-7f8858af71d8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] instance snapshotting [ 857.169765] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c72f26d8-b51b-417c-af61-21eeb3ed0629 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.191333] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3f20a6b-b900-48e1-bde5-e4630b5b5185 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.216126] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d9c956f-96c7-4e9f-8bf2-e0890adf91a5 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 857.216428] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-69d8c582-b637-4754-b204-142808bf0908 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.223893] env[62385]: DEBUG oslo_vmware.api [None req-1d9c956f-96c7-4e9f-8bf2-e0890adf91a5 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for the task: (returnval){ [ 857.223893] env[62385]: value = "task-1205914" [ 857.223893] env[62385]: _type = "Task" [ 857.223893] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.232452] env[62385]: DEBUG oslo_vmware.api [None req-1d9c956f-96c7-4e9f-8bf2-e0890adf91a5 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1205914, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.253550] env[62385]: INFO nova.compute.manager [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Took 43.77 seconds to build instance. [ 857.397023] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.588s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.397641] env[62385]: DEBUG nova.compute.manager [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 857.401190] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.779s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.403434] env[62385]: INFO nova.compute.claims [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 857.438421] env[62385]: DEBUG nova.network.neutron [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Updating instance_info_cache with network_info: [{"id": "2e49e027-be2a-4225-9a34-1e2f9cfb6022", "address": "fa:16:3e:cf:bd:fd", "network": {"id": "b0a3dea4-1925-4920-8a8e-9dc66733d1da", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1951021425-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ac0c8638f064daba042ded4de8c1cc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee20e439-fed9-490e-97dd-f3c886977ae1", "external-id": "nsx-vlan-transportzone-357", "segmentation_id": 357, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e49e027-be", "ovs_interfaceid": "2e49e027-be2a-4225-9a34-1e2f9cfb6022", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.557429] env[62385]: DEBUG nova.objects.instance [None req-fb0b1271-f283-4294-8a90-8402f992edc2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lazy-loading 'flavor' on Instance uuid 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 857.563247] env[62385]: DEBUG nova.network.neutron [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Successfully created port: 2fdc640e-c457-46d4-8a9d-fe1f934bd29d {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 857.583689] env[62385]: DEBUG oslo_vmware.api [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Task: {'id': task-1205913, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.706902] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d86e0c75-5a2a-4394-8269-7f8858af71d8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Creating Snapshot of the VM instance {{(pid=62385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 857.706902] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-f904ad70-2e71-4f85-b151-f93317352898 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.713040] env[62385]: DEBUG oslo_vmware.api [None req-d86e0c75-5a2a-4394-8269-7f8858af71d8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for the task: (returnval){ [ 857.713040] env[62385]: value = "task-1205915" [ 857.713040] env[62385]: _type = "Task" [ 857.713040] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.721392] env[62385]: DEBUG oslo_vmware.api [None req-d86e0c75-5a2a-4394-8269-7f8858af71d8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205915, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.732606] env[62385]: DEBUG oslo_vmware.api [None req-1d9c956f-96c7-4e9f-8bf2-e0890adf91a5 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1205914, 'name': PowerOnVM_Task, 'duration_secs': 0.450038} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.732882] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d9c956f-96c7-4e9f-8bf2-e0890adf91a5 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 857.733093] env[62385]: DEBUG nova.compute.manager [None req-1d9c956f-96c7-4e9f-8bf2-e0890adf91a5 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 857.733861] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecf4b9fa-bff5-4bbe-8c88-3623722df79e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.757560] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0acecd51-eaaa-47f0-aea1-1cfd353c8665 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Lock "5e5f43e8-ce45-45c7-a244-bfa724ea5e3c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 57.131s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.910535] env[62385]: DEBUG nova.compute.utils [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 857.916337] env[62385]: DEBUG nova.compute.manager [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 857.916687] env[62385]: DEBUG nova.network.neutron [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 857.941470] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Releasing lock "refresh_cache-b7a286a2-9a19-4ef1-8a46-6e2a04e22413" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.941863] env[62385]: DEBUG nova.compute.manager [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Instance network_info: |[{"id": "2e49e027-be2a-4225-9a34-1e2f9cfb6022", "address": "fa:16:3e:cf:bd:fd", "network": {"id": "b0a3dea4-1925-4920-8a8e-9dc66733d1da", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1951021425-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ac0c8638f064daba042ded4de8c1cc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee20e439-fed9-490e-97dd-f3c886977ae1", "external-id": "nsx-vlan-transportzone-357", "segmentation_id": 357, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e49e027-be", "ovs_interfaceid": "2e49e027-be2a-4225-9a34-1e2f9cfb6022", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 857.942602] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cf:bd:fd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ee20e439-fed9-490e-97dd-f3c886977ae1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2e49e027-be2a-4225-9a34-1e2f9cfb6022', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 857.950897] env[62385]: DEBUG oslo.service.loopingcall [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 857.951727] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 857.952035] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d34bf670-49cb-47a3-aa70-52126c7879e6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.972993] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 857.972993] env[62385]: value = "task-1205916" [ 857.972993] env[62385]: _type = "Task" [ 857.972993] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.980851] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205916, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.013889] env[62385]: DEBUG nova.policy [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e1200cdf75b04233b4c3f4626d4e27cf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aea48968d4464ac0ab01c998c059107d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 858.063432] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fb0b1271-f283-4294-8a90-8402f992edc2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "refresh_cache-9b9fd167-d843-4b9d-9e4f-dbde5d3d629e" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.063618] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fb0b1271-f283-4294-8a90-8402f992edc2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquired lock "refresh_cache-9b9fd167-d843-4b9d-9e4f-dbde5d3d629e" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.063795] env[62385]: DEBUG nova.network.neutron [None req-fb0b1271-f283-4294-8a90-8402f992edc2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 858.063976] env[62385]: DEBUG nova.objects.instance [None req-fb0b1271-f283-4294-8a90-8402f992edc2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lazy-loading 'info_cache' on Instance uuid 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 858.088219] env[62385]: DEBUG oslo_vmware.api [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Task: {'id': task-1205913, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.653597} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.088527] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 699a57a9-5a1a-4cd1-8449-723400d9caf4/699a57a9-5a1a-4cd1-8449-723400d9caf4.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 858.088778] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 858.089054] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dd813b24-9226-4efd-9b84-f642d087c3a4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.098783] env[62385]: DEBUG oslo_vmware.api [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Waiting for the task: (returnval){ [ 858.098783] env[62385]: value = "task-1205917" [ 858.098783] env[62385]: _type = "Task" [ 858.098783] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.113945] env[62385]: DEBUG oslo_vmware.api [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Task: {'id': task-1205917, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.226758] env[62385]: DEBUG oslo_vmware.api [None req-d86e0c75-5a2a-4394-8269-7f8858af71d8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205915, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.260843] env[62385]: DEBUG nova.compute.manager [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 858.264372] env[62385]: DEBUG nova.compute.manager [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 858.266556] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6591d0d-e738-458f-ad75-ba371b4bfecb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.310884] env[62385]: DEBUG nova.compute.manager [req-759371d5-d79a-43c3-886e-51fa0b86078d req-2d6b22e1-c6c1-40ec-84d6-a687b6b5aba9 service nova] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Received event network-changed-2e49e027-be2a-4225-9a34-1e2f9cfb6022 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 858.311076] env[62385]: DEBUG nova.compute.manager [req-759371d5-d79a-43c3-886e-51fa0b86078d req-2d6b22e1-c6c1-40ec-84d6-a687b6b5aba9 service nova] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Refreshing instance network info cache due to event network-changed-2e49e027-be2a-4225-9a34-1e2f9cfb6022. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 858.311302] env[62385]: DEBUG oslo_concurrency.lockutils [req-759371d5-d79a-43c3-886e-51fa0b86078d req-2d6b22e1-c6c1-40ec-84d6-a687b6b5aba9 service nova] Acquiring lock "refresh_cache-b7a286a2-9a19-4ef1-8a46-6e2a04e22413" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.311456] env[62385]: DEBUG oslo_concurrency.lockutils [req-759371d5-d79a-43c3-886e-51fa0b86078d req-2d6b22e1-c6c1-40ec-84d6-a687b6b5aba9 service nova] Acquired lock "refresh_cache-b7a286a2-9a19-4ef1-8a46-6e2a04e22413" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.311619] env[62385]: DEBUG nova.network.neutron [req-759371d5-d79a-43c3-886e-51fa0b86078d req-2d6b22e1-c6c1-40ec-84d6-a687b6b5aba9 service nova] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Refreshing network info cache for port 2e49e027-be2a-4225-9a34-1e2f9cfb6022 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 858.417044] env[62385]: DEBUG nova.compute.manager [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 858.474222] env[62385]: DEBUG nova.network.neutron [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Successfully created port: 7a94d167-2bed-4857-b8ce-bdc920b1982a {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 858.487876] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205916, 'name': CreateVM_Task} progress is 25%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.567553] env[62385]: DEBUG nova.objects.base [None req-fb0b1271-f283-4294-8a90-8402f992edc2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Object Instance<9b9fd167-d843-4b9d-9e4f-dbde5d3d629e> lazy-loaded attributes: flavor,info_cache {{(pid=62385) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 858.609025] env[62385]: DEBUG oslo_vmware.api [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Task: {'id': task-1205917, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071581} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.609311] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 858.610140] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae4558ea-ca42-4c59-b6cf-adb57e518c01 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.635838] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] 699a57a9-5a1a-4cd1-8449-723400d9caf4/699a57a9-5a1a-4cd1-8449-723400d9caf4.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 858.638554] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c268fea6-4f72-4330-8adc-88bc21e94b69 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.658633] env[62385]: DEBUG oslo_vmware.api [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Waiting for the task: (returnval){ [ 858.658633] env[62385]: value = "task-1205918" [ 858.658633] env[62385]: _type = "Task" [ 858.658633] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.670987] env[62385]: DEBUG oslo_vmware.api [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Task: {'id': task-1205918, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.725883] env[62385]: DEBUG oslo_vmware.api [None req-d86e0c75-5a2a-4394-8269-7f8858af71d8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205915, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.732243] env[62385]: INFO nova.compute.manager [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Rescuing [ 858.732429] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Acquiring lock "refresh_cache-5e5f43e8-ce45-45c7-a244-bfa724ea5e3c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.735019] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Acquired lock "refresh_cache-5e5f43e8-ce45-45c7-a244-bfa724ea5e3c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.735019] env[62385]: DEBUG nova.network.neutron [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 858.777949] env[62385]: INFO nova.compute.manager [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] instance snapshotting [ 858.783468] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86ee6ecd-33c6-49e5-80f6-9bcceb657f86 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.789678] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.808260] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d20a7de-5ad9-4d8a-b11d-530b9d6c82b9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.840596] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f517cf90-9ef5-4b87-b0ac-0ed6694e6fbf {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.849976] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0c1e6be-cea4-42a0-ad8a-763aeb6380c3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.883083] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52f9085c-90bd-4567-93e9-0459544c8700 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.890796] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aab4e72-6020-4e31-9b99-4dc751799787 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.906451] env[62385]: DEBUG nova.compute.provider_tree [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 858.990418] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205916, 'name': CreateVM_Task, 'duration_secs': 0.929814} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.990615] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 858.991425] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.991575] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.991923] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 858.992839] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-68d9befa-31c3-4254-8511-31ebad7e152d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.997645] env[62385]: DEBUG oslo_vmware.api [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 858.997645] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]526a507e-fc4a-67f0-8545-f63dabc0f10a" [ 858.997645] env[62385]: _type = "Task" [ 858.997645] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.007102] env[62385]: DEBUG oslo_vmware.api [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]526a507e-fc4a-67f0-8545-f63dabc0f10a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.081459] env[62385]: DEBUG nova.network.neutron [req-759371d5-d79a-43c3-886e-51fa0b86078d req-2d6b22e1-c6c1-40ec-84d6-a687b6b5aba9 service nova] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Updated VIF entry in instance network info cache for port 2e49e027-be2a-4225-9a34-1e2f9cfb6022. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 859.081860] env[62385]: DEBUG nova.network.neutron [req-759371d5-d79a-43c3-886e-51fa0b86078d req-2d6b22e1-c6c1-40ec-84d6-a687b6b5aba9 service nova] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Updating instance_info_cache with network_info: [{"id": "2e49e027-be2a-4225-9a34-1e2f9cfb6022", "address": "fa:16:3e:cf:bd:fd", "network": {"id": "b0a3dea4-1925-4920-8a8e-9dc66733d1da", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1951021425-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ac0c8638f064daba042ded4de8c1cc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee20e439-fed9-490e-97dd-f3c886977ae1", "external-id": "nsx-vlan-transportzone-357", "segmentation_id": 357, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e49e027-be", "ovs_interfaceid": "2e49e027-be2a-4225-9a34-1e2f9cfb6022", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.169632] env[62385]: DEBUG oslo_vmware.api [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Task: {'id': task-1205918, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.227855] env[62385]: DEBUG oslo_vmware.api [None req-d86e0c75-5a2a-4394-8269-7f8858af71d8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205915, 'name': CreateSnapshot_Task, 'duration_secs': 1.507924} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.228157] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d86e0c75-5a2a-4394-8269-7f8858af71d8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Created Snapshot of the VM instance {{(pid=62385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 859.229129] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ee6f3c1-f33d-41d2-b9b6-80e2f8bbbd2d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.321493] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Creating Snapshot of the VM instance {{(pid=62385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 859.324210] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-3e866cfd-e363-4908-b06d-5dacee7b69fa {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.332735] env[62385]: DEBUG oslo_vmware.api [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Waiting for the task: (returnval){ [ 859.332735] env[62385]: value = "task-1205919" [ 859.332735] env[62385]: _type = "Task" [ 859.332735] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.351383] env[62385]: DEBUG oslo_vmware.api [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205919, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.409286] env[62385]: DEBUG nova.scheduler.client.report [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 859.425799] env[62385]: DEBUG nova.compute.manager [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 859.448789] env[62385]: DEBUG nova.virt.hardware [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 859.448987] env[62385]: DEBUG nova.virt.hardware [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 859.449130] env[62385]: DEBUG nova.virt.hardware [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 859.449291] env[62385]: DEBUG nova.virt.hardware [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 859.449441] env[62385]: DEBUG nova.virt.hardware [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 859.449590] env[62385]: DEBUG nova.virt.hardware [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 859.449799] env[62385]: DEBUG nova.virt.hardware [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 859.450280] env[62385]: DEBUG nova.virt.hardware [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 859.450530] env[62385]: DEBUG nova.virt.hardware [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 859.450886] env[62385]: DEBUG nova.virt.hardware [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 859.451137] env[62385]: DEBUG nova.virt.hardware [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 859.452288] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e35777ff-c10f-4fbc-8779-f2cb4f6983e4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.461729] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31d8351c-70f8-49a4-9094-397da8d4d7a2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.468090] env[62385]: DEBUG nova.network.neutron [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Updating instance_info_cache with network_info: [{"id": "823867e4-d6b8-42d1-8597-3c27f6190ab1", "address": "fa:16:3e:56:b4:61", "network": {"id": "b563b70c-d3f1-4f82-97e9-defaf3f9d14b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-55206131-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "8eaa81326f084441a87ac05796958abb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa", "external-id": "nsx-vlan-transportzone-132", "segmentation_id": 132, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap823867e4-d6", "ovs_interfaceid": "823867e4-d6b8-42d1-8597-3c27f6190ab1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.508506] env[62385]: DEBUG oslo_vmware.api [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]526a507e-fc4a-67f0-8545-f63dabc0f10a, 'name': SearchDatastore_Task, 'duration_secs': 0.009729} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.508863] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.509282] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 859.509553] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.509715] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.509927] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 859.510230] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-98f56282-5b40-4c57-9377-1d02e3e129f1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.523825] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 859.524062] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 859.527497] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c057d25c-d3ba-471e-a846-9060de91a6bb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.533762] env[62385]: DEBUG oslo_vmware.api [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 859.533762] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52bc30d3-873a-b60b-e596-b5d21ab4a31b" [ 859.533762] env[62385]: _type = "Task" [ 859.533762] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.541456] env[62385]: DEBUG oslo_vmware.api [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52bc30d3-873a-b60b-e596-b5d21ab4a31b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.584880] env[62385]: DEBUG oslo_concurrency.lockutils [req-759371d5-d79a-43c3-886e-51fa0b86078d req-2d6b22e1-c6c1-40ec-84d6-a687b6b5aba9 service nova] Releasing lock "refresh_cache-b7a286a2-9a19-4ef1-8a46-6e2a04e22413" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.631097] env[62385]: DEBUG nova.network.neutron [None req-fb0b1271-f283-4294-8a90-8402f992edc2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Updating instance_info_cache with network_info: [{"id": "404c104c-258f-480e-bfe2-4ace4e83a43c", "address": "fa:16:3e:87:f8:2a", "network": {"id": "06017161-984c-4657-9639-4e6bdb6d4c09", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1072882119-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.208", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d45b5346eab44707b4d81eaaf12b0935", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39cd75b0-9ec7-48ed-b57f-34da0c573a60", "external-id": "nsx-vlan-transportzone-751", "segmentation_id": 751, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap404c104c-25", "ovs_interfaceid": "404c104c-258f-480e-bfe2-4ace4e83a43c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.671116] env[62385]: DEBUG oslo_vmware.api [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Task: {'id': task-1205918, 'name': ReconfigVM_Task, 'duration_secs': 0.554836} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.671457] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Reconfigured VM instance instance-0000004b to attach disk [datastore1] 699a57a9-5a1a-4cd1-8449-723400d9caf4/699a57a9-5a1a-4cd1-8449-723400d9caf4.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 859.672400] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-45c4aab7-32b9-453b-860c-9ac8b7c1680c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.681897] env[62385]: DEBUG oslo_vmware.api [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Waiting for the task: (returnval){ [ 859.681897] env[62385]: value = "task-1205920" [ 859.681897] env[62385]: _type = "Task" [ 859.681897] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.692042] env[62385]: DEBUG oslo_vmware.api [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Task: {'id': task-1205920, 'name': Rename_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.752165] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d86e0c75-5a2a-4394-8269-7f8858af71d8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Creating linked-clone VM from snapshot {{(pid=62385) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 859.752556] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-1b8b29b2-0845-46b9-97b4-eba275188cbb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.774831] env[62385]: DEBUG oslo_vmware.api [None req-d86e0c75-5a2a-4394-8269-7f8858af71d8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for the task: (returnval){ [ 859.774831] env[62385]: value = "task-1205921" [ 859.774831] env[62385]: _type = "Task" [ 859.774831] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.777233] env[62385]: DEBUG nova.network.neutron [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Successfully updated port: 2572a945-d7b9-47e5-8641-3718434e80ca {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 859.790766] env[62385]: DEBUG oslo_vmware.api [None req-d86e0c75-5a2a-4394-8269-7f8858af71d8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205921, 'name': CloneVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.844913] env[62385]: DEBUG oslo_vmware.api [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205919, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.918023] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.514s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.918023] env[62385]: DEBUG nova.compute.manager [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 859.919269] env[62385]: DEBUG oslo_concurrency.lockutils [None req-71e9a3c9-f8fb-42f2-a138-6fdd3a07054e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.255s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.919971] env[62385]: DEBUG nova.objects.instance [None req-71e9a3c9-f8fb-42f2-a138-6fdd3a07054e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lazy-loading 'resources' on Instance uuid 55d62bba-ff30-42bf-b1cb-7567988a9361 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 859.972949] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Releasing lock "refresh_cache-5e5f43e8-ce45-45c7-a244-bfa724ea5e3c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.044968] env[62385]: DEBUG oslo_vmware.api [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52bc30d3-873a-b60b-e596-b5d21ab4a31b, 'name': SearchDatastore_Task, 'duration_secs': 0.012355} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.045850] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd67365b-9499-429f-aef8-489881113e2e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.051828] env[62385]: DEBUG oslo_vmware.api [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 860.051828] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5227ae22-73a9-a066-2849-7935e220e8a1" [ 860.051828] env[62385]: _type = "Task" [ 860.051828] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.060506] env[62385]: DEBUG oslo_vmware.api [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5227ae22-73a9-a066-2849-7935e220e8a1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.138932] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fb0b1271-f283-4294-8a90-8402f992edc2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Releasing lock "refresh_cache-9b9fd167-d843-4b9d-9e4f-dbde5d3d629e" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.193061] env[62385]: DEBUG oslo_vmware.api [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Task: {'id': task-1205920, 'name': Rename_Task, 'duration_secs': 0.171275} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.193061] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 860.193061] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fd83f335-ea99-4c65-ba39-75df9c84fb94 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.202946] env[62385]: DEBUG oslo_vmware.api [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Waiting for the task: (returnval){ [ 860.202946] env[62385]: value = "task-1205922" [ 860.202946] env[62385]: _type = "Task" [ 860.202946] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.207256] env[62385]: DEBUG oslo_vmware.api [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Task: {'id': task-1205922, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.290427] env[62385]: DEBUG oslo_vmware.api [None req-d86e0c75-5a2a-4394-8269-7f8858af71d8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205921, 'name': CloneVM_Task} progress is 94%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.344119] env[62385]: DEBUG oslo_vmware.api [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205919, 'name': CreateSnapshot_Task, 'duration_secs': 0.849595} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.344397] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Created Snapshot of the VM instance {{(pid=62385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 860.345146] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0077391d-6c4b-4c7f-85b0-f720e6f7619b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.391300] env[62385]: DEBUG nova.compute.manager [req-09f9fea8-cbd0-4019-820e-cd24e1062a9b req-8c6af49c-a3bb-4c91-a9f8-0b018206f380 service nova] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Received event network-vif-plugged-2572a945-d7b9-47e5-8641-3718434e80ca {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 860.391599] env[62385]: DEBUG oslo_concurrency.lockutils [req-09f9fea8-cbd0-4019-820e-cd24e1062a9b req-8c6af49c-a3bb-4c91-a9f8-0b018206f380 service nova] Acquiring lock "fbc71ada-d3b5-46f6-90a9-489c118d5126-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.391741] env[62385]: DEBUG oslo_concurrency.lockutils [req-09f9fea8-cbd0-4019-820e-cd24e1062a9b req-8c6af49c-a3bb-4c91-a9f8-0b018206f380 service nova] Lock "fbc71ada-d3b5-46f6-90a9-489c118d5126-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.391914] env[62385]: DEBUG oslo_concurrency.lockutils [req-09f9fea8-cbd0-4019-820e-cd24e1062a9b req-8c6af49c-a3bb-4c91-a9f8-0b018206f380 service nova] Lock "fbc71ada-d3b5-46f6-90a9-489c118d5126-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.393863] env[62385]: DEBUG nova.compute.manager [req-09f9fea8-cbd0-4019-820e-cd24e1062a9b req-8c6af49c-a3bb-4c91-a9f8-0b018206f380 service nova] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] No waiting events found dispatching network-vif-plugged-2572a945-d7b9-47e5-8641-3718434e80ca {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 860.393863] env[62385]: WARNING nova.compute.manager [req-09f9fea8-cbd0-4019-820e-cd24e1062a9b req-8c6af49c-a3bb-4c91-a9f8-0b018206f380 service nova] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Received unexpected event network-vif-plugged-2572a945-d7b9-47e5-8641-3718434e80ca for instance with vm_state building and task_state spawning. [ 860.393863] env[62385]: DEBUG nova.compute.manager [req-09f9fea8-cbd0-4019-820e-cd24e1062a9b req-8c6af49c-a3bb-4c91-a9f8-0b018206f380 service nova] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Received event network-changed-2572a945-d7b9-47e5-8641-3718434e80ca {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 860.393863] env[62385]: DEBUG nova.compute.manager [req-09f9fea8-cbd0-4019-820e-cd24e1062a9b req-8c6af49c-a3bb-4c91-a9f8-0b018206f380 service nova] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Refreshing instance network info cache due to event network-changed-2572a945-d7b9-47e5-8641-3718434e80ca. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 860.394585] env[62385]: DEBUG oslo_concurrency.lockutils [req-09f9fea8-cbd0-4019-820e-cd24e1062a9b req-8c6af49c-a3bb-4c91-a9f8-0b018206f380 service nova] Acquiring lock "refresh_cache-fbc71ada-d3b5-46f6-90a9-489c118d5126" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.397872] env[62385]: DEBUG oslo_concurrency.lockutils [req-09f9fea8-cbd0-4019-820e-cd24e1062a9b req-8c6af49c-a3bb-4c91-a9f8-0b018206f380 service nova] Acquired lock "refresh_cache-fbc71ada-d3b5-46f6-90a9-489c118d5126" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.397872] env[62385]: DEBUG nova.network.neutron [req-09f9fea8-cbd0-4019-820e-cd24e1062a9b req-8c6af49c-a3bb-4c91-a9f8-0b018206f380 service nova] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Refreshing network info cache for port 2572a945-d7b9-47e5-8641-3718434e80ca {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 860.426416] env[62385]: DEBUG nova.compute.utils [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 860.435801] env[62385]: DEBUG nova.compute.manager [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 860.435801] env[62385]: DEBUG nova.network.neutron [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 860.507899] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 860.508270] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-aa848a3c-44f7-41fc-86fa-6239c8fdf579 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.517333] env[62385]: DEBUG oslo_vmware.api [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for the task: (returnval){ [ 860.517333] env[62385]: value = "task-1205923" [ 860.517333] env[62385]: _type = "Task" [ 860.517333] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.532088] env[62385]: DEBUG oslo_vmware.api [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205923, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.573260] env[62385]: DEBUG oslo_vmware.api [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5227ae22-73a9-a066-2849-7935e220e8a1, 'name': SearchDatastore_Task, 'duration_secs': 0.015884} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.573573] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.573903] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] b7a286a2-9a19-4ef1-8a46-6e2a04e22413/b7a286a2-9a19-4ef1-8a46-6e2a04e22413.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 860.574226] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3422805c-f7ed-47ce-9244-e823d806f037 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.582657] env[62385]: DEBUG oslo_vmware.api [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 860.582657] env[62385]: value = "task-1205924" [ 860.582657] env[62385]: _type = "Task" [ 860.582657] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.595720] env[62385]: DEBUG oslo_vmware.api [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205924, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.642637] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb0b1271-f283-4294-8a90-8402f992edc2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 860.645502] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ffdaf9e4-5e05-4cd2-83e3-a07632ec88a1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.655961] env[62385]: DEBUG oslo_vmware.api [None req-fb0b1271-f283-4294-8a90-8402f992edc2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 860.655961] env[62385]: value = "task-1205925" [ 860.655961] env[62385]: _type = "Task" [ 860.655961] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.667889] env[62385]: DEBUG oslo_vmware.api [None req-fb0b1271-f283-4294-8a90-8402f992edc2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1205925, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.713072] env[62385]: DEBUG oslo_vmware.api [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Task: {'id': task-1205922, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.790550] env[62385]: DEBUG oslo_vmware.api [None req-d86e0c75-5a2a-4394-8269-7f8858af71d8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205921, 'name': CloneVM_Task} progress is 94%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.794127] env[62385]: DEBUG nova.policy [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '12f9521f9c9b4c6b9afc4973be32c5b7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '731b6700ce5b43369b40165545f04950', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 860.866092] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Creating linked-clone VM from snapshot {{(pid=62385) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 860.866417] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-fc89c681-7830-41a9-a40e-0a0f4aba57d8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.873396] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1152ca3d-374d-41cd-b229-61583ef1951c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.887999] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f60bca0d-a0b5-40a9-aa79-7baf8161b769 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.893364] env[62385]: DEBUG oslo_vmware.api [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Waiting for the task: (returnval){ [ 860.893364] env[62385]: value = "task-1205926" [ 860.893364] env[62385]: _type = "Task" [ 860.893364] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.940662] env[62385]: DEBUG nova.compute.manager [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 860.946604] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1544e23-8c77-4246-87f7-b7fa448ef981 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.957439] env[62385]: DEBUG oslo_vmware.api [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205926, 'name': CloneVM_Task} progress is 12%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.964683] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa1498bf-ec4c-40e3-b4c4-2a495c203c1f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.986226] env[62385]: DEBUG nova.compute.provider_tree [None req-71e9a3c9-f8fb-42f2-a138-6fdd3a07054e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 861.018322] env[62385]: DEBUG nova.network.neutron [req-09f9fea8-cbd0-4019-820e-cd24e1062a9b req-8c6af49c-a3bb-4c91-a9f8-0b018206f380 service nova] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 861.033695] env[62385]: DEBUG oslo_vmware.api [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205923, 'name': PowerOffVM_Task, 'duration_secs': 0.354372} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.034101] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 861.035276] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f92ff535-cd77-4f0b-b637-11c2bef1783f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.065357] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ffc5274-e8ed-403d-9b2d-8ed3618faf5a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.096989] env[62385]: DEBUG oslo_vmware.api [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205924, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.115771] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 861.116260] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-341eb3d3-998e-4c20-b23e-59cb039d1fa0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.120341] env[62385]: DEBUG nova.network.neutron [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Successfully updated port: 7a94d167-2bed-4857-b8ce-bdc920b1982a {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 861.126439] env[62385]: DEBUG oslo_vmware.api [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for the task: (returnval){ [ 861.126439] env[62385]: value = "task-1205927" [ 861.126439] env[62385]: _type = "Task" [ 861.126439] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.137460] env[62385]: DEBUG oslo_vmware.api [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205927, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.138553] env[62385]: DEBUG nova.network.neutron [req-09f9fea8-cbd0-4019-820e-cd24e1062a9b req-8c6af49c-a3bb-4c91-a9f8-0b018206f380 service nova] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.172819] env[62385]: DEBUG oslo_vmware.api [None req-fb0b1271-f283-4294-8a90-8402f992edc2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1205925, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.214572] env[62385]: DEBUG oslo_vmware.api [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Task: {'id': task-1205922, 'name': PowerOnVM_Task, 'duration_secs': 0.781224} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.214867] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 861.215106] env[62385]: INFO nova.compute.manager [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Took 9.39 seconds to spawn the instance on the hypervisor. [ 861.215271] env[62385]: DEBUG nova.compute.manager [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 861.216632] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25a83a85-bf89-410e-a781-4377db1e9e8f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.292712] env[62385]: DEBUG oslo_vmware.api [None req-d86e0c75-5a2a-4394-8269-7f8858af71d8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205921, 'name': CloneVM_Task} progress is 94%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.405013] env[62385]: DEBUG oslo_vmware.api [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205926, 'name': CloneVM_Task} progress is 94%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.489587] env[62385]: DEBUG nova.scheduler.client.report [None req-71e9a3c9-f8fb-42f2-a138-6fdd3a07054e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 861.596773] env[62385]: DEBUG oslo_vmware.api [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205924, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.679701} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.597186] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] b7a286a2-9a19-4ef1-8a46-6e2a04e22413/b7a286a2-9a19-4ef1-8a46-6e2a04e22413.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 861.597534] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 861.597902] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f95ab852-be98-448b-bf78-b585ce79359b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.611492] env[62385]: DEBUG oslo_vmware.api [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 861.611492] env[62385]: value = "task-1205928" [ 861.611492] env[62385]: _type = "Task" [ 861.611492] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.624293] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "refresh_cache-fe0624a9-09a8-498b-bb3c-fda6cab92341" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.624508] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquired lock "refresh_cache-fe0624a9-09a8-498b-bb3c-fda6cab92341" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.624695] env[62385]: DEBUG nova.network.neutron [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 861.626316] env[62385]: DEBUG oslo_vmware.api [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205928, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.641195] env[62385]: DEBUG oslo_concurrency.lockutils [req-09f9fea8-cbd0-4019-820e-cd24e1062a9b req-8c6af49c-a3bb-4c91-a9f8-0b018206f380 service nova] Releasing lock "refresh_cache-fbc71ada-d3b5-46f6-90a9-489c118d5126" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.641649] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] VM already powered off {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 861.641858] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 861.642117] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.642276] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.642460] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 861.642988] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f26dd727-1c2c-4c2d-8262-ce41b856f9b5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.657539] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 861.657742] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 861.662076] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-00f2e10b-5d21-4f69-88aa-fea5e41c88e9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.674691] env[62385]: DEBUG oslo_vmware.api [None req-fb0b1271-f283-4294-8a90-8402f992edc2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1205925, 'name': PowerOnVM_Task, 'duration_secs': 0.700219} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.675020] env[62385]: DEBUG oslo_vmware.api [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for the task: (returnval){ [ 861.675020] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]527aa42a-0d38-b35d-f5da-8d584fe149a0" [ 861.675020] env[62385]: _type = "Task" [ 861.675020] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.675290] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb0b1271-f283-4294-8a90-8402f992edc2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 861.675480] env[62385]: DEBUG nova.compute.manager [None req-fb0b1271-f283-4294-8a90-8402f992edc2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 861.676409] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57a5e0f0-d27f-4532-8e66-05c40799e357 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.684855] env[62385]: DEBUG nova.network.neutron [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Successfully created port: fa9f8b71-802a-46a3-a93b-20ee02c6adcb {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 861.700209] env[62385]: DEBUG oslo_vmware.api [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]527aa42a-0d38-b35d-f5da-8d584fe149a0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.737576] env[62385]: INFO nova.compute.manager [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Took 34.45 seconds to build instance. [ 861.790467] env[62385]: DEBUG oslo_vmware.api [None req-d86e0c75-5a2a-4394-8269-7f8858af71d8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205921, 'name': CloneVM_Task, 'duration_secs': 1.878252} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.790739] env[62385]: INFO nova.virt.vmwareapi.vmops [None req-d86e0c75-5a2a-4394-8269-7f8858af71d8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Created linked-clone VM from snapshot [ 861.791509] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4bb6313-f75f-4f80-a11e-3f9086980ccb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.800014] env[62385]: DEBUG nova.virt.vmwareapi.images [None req-d86e0c75-5a2a-4394-8269-7f8858af71d8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Uploading image e3323aab-9833-47a5-ab60-9095f41291a9 {{(pid=62385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 861.818532] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d86e0c75-5a2a-4394-8269-7f8858af71d8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Destroying the VM {{(pid=62385) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 861.818532] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-36bb5a70-1095-41e1-ac67-ee0f5aec9987 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.825905] env[62385]: DEBUG oslo_vmware.api [None req-d86e0c75-5a2a-4394-8269-7f8858af71d8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for the task: (returnval){ [ 861.825905] env[62385]: value = "task-1205929" [ 861.825905] env[62385]: _type = "Task" [ 861.825905] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.835059] env[62385]: DEBUG oslo_vmware.api [None req-d86e0c75-5a2a-4394-8269-7f8858af71d8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205929, 'name': Destroy_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.906534] env[62385]: DEBUG oslo_vmware.api [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205926, 'name': CloneVM_Task} progress is 94%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.952455] env[62385]: DEBUG nova.compute.manager [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 861.980274] env[62385]: DEBUG nova.virt.hardware [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 861.980542] env[62385]: DEBUG nova.virt.hardware [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 861.980709] env[62385]: DEBUG nova.virt.hardware [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 861.980895] env[62385]: DEBUG nova.virt.hardware [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 861.981073] env[62385]: DEBUG nova.virt.hardware [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 861.981291] env[62385]: DEBUG nova.virt.hardware [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 861.981509] env[62385]: DEBUG nova.virt.hardware [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 861.981675] env[62385]: DEBUG nova.virt.hardware [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 861.981848] env[62385]: DEBUG nova.virt.hardware [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 861.982031] env[62385]: DEBUG nova.virt.hardware [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 861.982254] env[62385]: DEBUG nova.virt.hardware [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 861.983236] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bf8af05-e040-4928-b71c-f0c0ab09270a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.991536] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a22b039f-1a0d-4b76-bc39-7868a59240a2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.996034] env[62385]: DEBUG oslo_concurrency.lockutils [None req-71e9a3c9-f8fb-42f2-a138-6fdd3a07054e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.077s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.998071] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 14.694s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.998268] env[62385]: DEBUG nova.objects.instance [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62385) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 862.019154] env[62385]: INFO nova.scheduler.client.report [None req-71e9a3c9-f8fb-42f2-a138-6fdd3a07054e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Deleted allocations for instance 55d62bba-ff30-42bf-b1cb-7567988a9361 [ 862.122699] env[62385]: DEBUG oslo_vmware.api [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205928, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072239} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.123106] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 862.123696] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d015450-b055-4dcb-862d-d0f833e084f3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.151155] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] b7a286a2-9a19-4ef1-8a46-6e2a04e22413/b7a286a2-9a19-4ef1-8a46-6e2a04e22413.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 862.154834] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1c30dd34-5314-46e9-b8d7-a11cd9c0710a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.181570] env[62385]: DEBUG oslo_vmware.api [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 862.181570] env[62385]: value = "task-1205930" [ 862.181570] env[62385]: _type = "Task" [ 862.181570] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.189168] env[62385]: DEBUG oslo_vmware.api [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]527aa42a-0d38-b35d-f5da-8d584fe149a0, 'name': SearchDatastore_Task, 'duration_secs': 0.038518} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.190846] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d654647f-aaf7-4011-b5ea-d03bb5cb39b9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.195969] env[62385]: DEBUG oslo_vmware.api [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205930, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.201354] env[62385]: DEBUG oslo_vmware.api [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for the task: (returnval){ [ 862.201354] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5227025c-9ba1-5313-ba41-6d4f156e4410" [ 862.201354] env[62385]: _type = "Task" [ 862.201354] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.203306] env[62385]: DEBUG nova.network.neutron [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 862.212592] env[62385]: DEBUG oslo_vmware.api [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5227025c-9ba1-5313-ba41-6d4f156e4410, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.240018] env[62385]: DEBUG oslo_concurrency.lockutils [None req-63c81ab8-67bd-4b70-9176-51d7f754cb05 tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Lock "699a57a9-5a1a-4cd1-8449-723400d9caf4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.965s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.337942] env[62385]: DEBUG oslo_vmware.api [None req-d86e0c75-5a2a-4394-8269-7f8858af71d8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205929, 'name': Destroy_Task} progress is 33%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.371911] env[62385]: DEBUG nova.network.neutron [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Updating instance_info_cache with network_info: [{"id": "7a94d167-2bed-4857-b8ce-bdc920b1982a", "address": "fa:16:3e:4b:17:c5", "network": {"id": "ff21dd88-fa02-4bd6-8009-980ec6e72758", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1065789169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aea48968d4464ac0ab01c998c059107d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3ff3baee-99ce-4b51-ae98-efc6163aaab3", "external-id": "nsx-vlan-transportzone-574", "segmentation_id": 574, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7a94d167-2b", "ovs_interfaceid": "7a94d167-2bed-4857-b8ce-bdc920b1982a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.407188] env[62385]: DEBUG oslo_vmware.api [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205926, 'name': CloneVM_Task} progress is 94%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.424927] env[62385]: DEBUG nova.compute.manager [req-89b7aad2-9596-4553-a4b0-4c2d5ffcbe7c req-cb9061aa-0f43-4ac9-b76b-f19679cb9da5 service nova] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Received event network-vif-plugged-7a94d167-2bed-4857-b8ce-bdc920b1982a {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 862.425185] env[62385]: DEBUG oslo_concurrency.lockutils [req-89b7aad2-9596-4553-a4b0-4c2d5ffcbe7c req-cb9061aa-0f43-4ac9-b76b-f19679cb9da5 service nova] Acquiring lock "fe0624a9-09a8-498b-bb3c-fda6cab92341-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.425439] env[62385]: DEBUG oslo_concurrency.lockutils [req-89b7aad2-9596-4553-a4b0-4c2d5ffcbe7c req-cb9061aa-0f43-4ac9-b76b-f19679cb9da5 service nova] Lock "fe0624a9-09a8-498b-bb3c-fda6cab92341-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.425610] env[62385]: DEBUG oslo_concurrency.lockutils [req-89b7aad2-9596-4553-a4b0-4c2d5ffcbe7c req-cb9061aa-0f43-4ac9-b76b-f19679cb9da5 service nova] Lock "fe0624a9-09a8-498b-bb3c-fda6cab92341-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.425787] env[62385]: DEBUG nova.compute.manager [req-89b7aad2-9596-4553-a4b0-4c2d5ffcbe7c req-cb9061aa-0f43-4ac9-b76b-f19679cb9da5 service nova] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] No waiting events found dispatching network-vif-plugged-7a94d167-2bed-4857-b8ce-bdc920b1982a {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 862.425977] env[62385]: WARNING nova.compute.manager [req-89b7aad2-9596-4553-a4b0-4c2d5ffcbe7c req-cb9061aa-0f43-4ac9-b76b-f19679cb9da5 service nova] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Received unexpected event network-vif-plugged-7a94d167-2bed-4857-b8ce-bdc920b1982a for instance with vm_state building and task_state spawning. [ 862.426190] env[62385]: DEBUG nova.compute.manager [req-89b7aad2-9596-4553-a4b0-4c2d5ffcbe7c req-cb9061aa-0f43-4ac9-b76b-f19679cb9da5 service nova] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Received event network-changed-7a94d167-2bed-4857-b8ce-bdc920b1982a {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 862.426350] env[62385]: DEBUG nova.compute.manager [req-89b7aad2-9596-4553-a4b0-4c2d5ffcbe7c req-cb9061aa-0f43-4ac9-b76b-f19679cb9da5 service nova] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Refreshing instance network info cache due to event network-changed-7a94d167-2bed-4857-b8ce-bdc920b1982a. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 862.426520] env[62385]: DEBUG oslo_concurrency.lockutils [req-89b7aad2-9596-4553-a4b0-4c2d5ffcbe7c req-cb9061aa-0f43-4ac9-b76b-f19679cb9da5 service nova] Acquiring lock "refresh_cache-fe0624a9-09a8-498b-bb3c-fda6cab92341" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.528334] env[62385]: DEBUG oslo_concurrency.lockutils [None req-71e9a3c9-f8fb-42f2-a138-6fdd3a07054e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "55d62bba-ff30-42bf-b1cb-7567988a9361" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.976s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.695109] env[62385]: DEBUG oslo_vmware.api [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205930, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.723512] env[62385]: DEBUG oslo_vmware.api [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5227025c-9ba1-5313-ba41-6d4f156e4410, 'name': SearchDatastore_Task, 'duration_secs': 0.012886} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.723960] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.724341] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c/c0abbb8d-77e7-4f22-a256-0faf0a781109-rescue.vmdk. {{(pid=62385) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 862.724661] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0cac58d0-deb7-43b9-af54-9a457cc0fe84 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.734812] env[62385]: DEBUG oslo_vmware.api [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for the task: (returnval){ [ 862.734812] env[62385]: value = "task-1205931" [ 862.734812] env[62385]: _type = "Task" [ 862.734812] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.750432] env[62385]: DEBUG oslo_vmware.api [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205931, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.837352] env[62385]: DEBUG oslo_vmware.api [None req-d86e0c75-5a2a-4394-8269-7f8858af71d8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205929, 'name': Destroy_Task, 'duration_secs': 0.879678} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.837743] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-d86e0c75-5a2a-4394-8269-7f8858af71d8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Destroyed the VM [ 862.838083] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d86e0c75-5a2a-4394-8269-7f8858af71d8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Deleting Snapshot of the VM instance {{(pid=62385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 862.838384] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-a4fe7490-fe3c-4f82-b884-b52d02ee84eb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.845721] env[62385]: DEBUG oslo_vmware.api [None req-d86e0c75-5a2a-4394-8269-7f8858af71d8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for the task: (returnval){ [ 862.845721] env[62385]: value = "task-1205932" [ 862.845721] env[62385]: _type = "Task" [ 862.845721] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.855908] env[62385]: DEBUG oslo_vmware.api [None req-d86e0c75-5a2a-4394-8269-7f8858af71d8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205932, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.874757] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Releasing lock "refresh_cache-fe0624a9-09a8-498b-bb3c-fda6cab92341" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.875192] env[62385]: DEBUG nova.compute.manager [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Instance network_info: |[{"id": "7a94d167-2bed-4857-b8ce-bdc920b1982a", "address": "fa:16:3e:4b:17:c5", "network": {"id": "ff21dd88-fa02-4bd6-8009-980ec6e72758", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1065789169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aea48968d4464ac0ab01c998c059107d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3ff3baee-99ce-4b51-ae98-efc6163aaab3", "external-id": "nsx-vlan-transportzone-574", "segmentation_id": 574, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7a94d167-2b", "ovs_interfaceid": "7a94d167-2bed-4857-b8ce-bdc920b1982a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 862.875562] env[62385]: DEBUG oslo_concurrency.lockutils [req-89b7aad2-9596-4553-a4b0-4c2d5ffcbe7c req-cb9061aa-0f43-4ac9-b76b-f19679cb9da5 service nova] Acquired lock "refresh_cache-fe0624a9-09a8-498b-bb3c-fda6cab92341" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.875802] env[62385]: DEBUG nova.network.neutron [req-89b7aad2-9596-4553-a4b0-4c2d5ffcbe7c req-cb9061aa-0f43-4ac9-b76b-f19679cb9da5 service nova] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Refreshing network info cache for port 7a94d167-2bed-4857-b8ce-bdc920b1982a {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 862.877380] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4b:17:c5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3ff3baee-99ce-4b51-ae98-efc6163aaab3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7a94d167-2bed-4857-b8ce-bdc920b1982a', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 862.887858] env[62385]: DEBUG oslo.service.loopingcall [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 862.891771] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 862.892956] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-52adcb7b-fc1b-45b6-a42a-1d2d8a8340b2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.925588] env[62385]: DEBUG oslo_vmware.api [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205926, 'name': CloneVM_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.927849] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 862.927849] env[62385]: value = "task-1205933" [ 862.927849] env[62385]: _type = "Task" [ 862.927849] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.937880] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205933, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.008174] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bc01508f-5cd8-4335-a3fd-cb10885fb5b0 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.010498] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5dfcbc54-c64f-4274-aed3-ea9558593cc5 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.429s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.010864] env[62385]: DEBUG nova.objects.instance [None req-5dfcbc54-c64f-4274-aed3-ea9558593cc5 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lazy-loading 'resources' on Instance uuid 0d823fcc-cbba-41f6-a47c-2f57d3e79948 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 863.156710] env[62385]: DEBUG nova.network.neutron [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Successfully updated port: 6820d526-f9c3-43cc-91db-538e9c3bd249 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 863.171053] env[62385]: DEBUG nova.network.neutron [req-89b7aad2-9596-4553-a4b0-4c2d5ffcbe7c req-cb9061aa-0f43-4ac9-b76b-f19679cb9da5 service nova] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Updated VIF entry in instance network info cache for port 7a94d167-2bed-4857-b8ce-bdc920b1982a. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 863.171441] env[62385]: DEBUG nova.network.neutron [req-89b7aad2-9596-4553-a4b0-4c2d5ffcbe7c req-cb9061aa-0f43-4ac9-b76b-f19679cb9da5 service nova] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Updating instance_info_cache with network_info: [{"id": "7a94d167-2bed-4857-b8ce-bdc920b1982a", "address": "fa:16:3e:4b:17:c5", "network": {"id": "ff21dd88-fa02-4bd6-8009-980ec6e72758", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1065789169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aea48968d4464ac0ab01c998c059107d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3ff3baee-99ce-4b51-ae98-efc6163aaab3", "external-id": "nsx-vlan-transportzone-574", "segmentation_id": 574, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7a94d167-2b", "ovs_interfaceid": "7a94d167-2bed-4857-b8ce-bdc920b1982a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.195865] env[62385]: DEBUG oslo_vmware.api [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205930, 'name': ReconfigVM_Task, 'duration_secs': 0.715839} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.196145] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Reconfigured VM instance instance-0000004c to attach disk [datastore1] b7a286a2-9a19-4ef1-8a46-6e2a04e22413/b7a286a2-9a19-4ef1-8a46-6e2a04e22413.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 863.196909] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-729da208-d3ba-4d38-9671-2f72b6cd1f0e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.206489] env[62385]: DEBUG oslo_vmware.api [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 863.206489] env[62385]: value = "task-1205934" [ 863.206489] env[62385]: _type = "Task" [ 863.206489] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.217904] env[62385]: DEBUG oslo_vmware.api [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205934, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.250331] env[62385]: DEBUG oslo_vmware.api [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205931, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.359209] env[62385]: DEBUG oslo_vmware.api [None req-d86e0c75-5a2a-4394-8269-7f8858af71d8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205932, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.427473] env[62385]: DEBUG oslo_vmware.api [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205926, 'name': CloneVM_Task, 'duration_secs': 2.113207} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.427801] env[62385]: INFO nova.virt.vmwareapi.vmops [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Created linked-clone VM from snapshot [ 863.428623] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93f55862-ea1b-4106-aca9-0914e6376690 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.440866] env[62385]: DEBUG nova.virt.vmwareapi.images [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Uploading image a204cf12-3e93-4857-af15-299e714bfa66 {{(pid=62385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 863.446626] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205933, 'name': CreateVM_Task, 'duration_secs': 0.466601} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.446828] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 863.447548] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.447763] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.448180] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 863.449105] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-25ae0a59-c957-45d6-9b12-bc86913ec6bb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.454335] env[62385]: DEBUG oslo_vmware.api [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 863.454335] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52666ac2-4eff-e013-d15a-dc2b5d6dc2ac" [ 863.454335] env[62385]: _type = "Task" [ 863.454335] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.465738] env[62385]: DEBUG oslo_vmware.api [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52666ac2-4eff-e013-d15a-dc2b5d6dc2ac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.467803] env[62385]: DEBUG oslo_vmware.rw_handles [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 863.467803] env[62385]: value = "vm-261230" [ 863.467803] env[62385]: _type = "VirtualMachine" [ 863.467803] env[62385]: }. {{(pid=62385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 863.468102] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-87020ac6-fc42-455e-bf0e-050df54a1802 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.475211] env[62385]: DEBUG oslo_vmware.rw_handles [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Lease: (returnval){ [ 863.475211] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5267ea8f-cbbc-8b54-2b2b-ac7c53e52d7b" [ 863.475211] env[62385]: _type = "HttpNfcLease" [ 863.475211] env[62385]: } obtained for exporting VM: (result){ [ 863.475211] env[62385]: value = "vm-261230" [ 863.475211] env[62385]: _type = "VirtualMachine" [ 863.475211] env[62385]: }. {{(pid=62385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 863.475590] env[62385]: DEBUG oslo_vmware.api [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Waiting for the lease: (returnval){ [ 863.475590] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5267ea8f-cbbc-8b54-2b2b-ac7c53e52d7b" [ 863.475590] env[62385]: _type = "HttpNfcLease" [ 863.475590] env[62385]: } to be ready. {{(pid=62385) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 863.482626] env[62385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 863.482626] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5267ea8f-cbbc-8b54-2b2b-ac7c53e52d7b" [ 863.482626] env[62385]: _type = "HttpNfcLease" [ 863.482626] env[62385]: } is initializing. {{(pid=62385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 863.674267] env[62385]: DEBUG oslo_concurrency.lockutils [req-89b7aad2-9596-4553-a4b0-4c2d5ffcbe7c req-cb9061aa-0f43-4ac9-b76b-f19679cb9da5 service nova] Releasing lock "refresh_cache-fe0624a9-09a8-498b-bb3c-fda6cab92341" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.725907] env[62385]: DEBUG oslo_vmware.api [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205934, 'name': Rename_Task, 'duration_secs': 0.259258} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.726318] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 863.727119] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-645f37d6-2d5b-44f4-9232-2709e6ca83b9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.735402] env[62385]: DEBUG oslo_vmware.api [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 863.735402] env[62385]: value = "task-1205936" [ 863.735402] env[62385]: _type = "Task" [ 863.735402] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.747819] env[62385]: DEBUG oslo_vmware.api [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205936, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.753627] env[62385]: DEBUG oslo_vmware.api [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205931, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.601442} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.755042] env[62385]: INFO nova.virt.vmwareapi.ds_util [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c/c0abbb8d-77e7-4f22-a256-0faf0a781109-rescue.vmdk. [ 863.755042] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-696f8024-3a97-4ac0-891a-8f8aced76249 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.783178] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c/c0abbb8d-77e7-4f22-a256-0faf0a781109-rescue.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 863.786182] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cc65e4c0-185a-4ff7-81a7-cdfd632da93c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.800251] env[62385]: DEBUG nova.network.neutron [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Successfully updated port: fa9f8b71-802a-46a3-a93b-20ee02c6adcb {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 863.809241] env[62385]: DEBUG oslo_vmware.api [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for the task: (returnval){ [ 863.809241] env[62385]: value = "task-1205937" [ 863.809241] env[62385]: _type = "Task" [ 863.809241] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.823183] env[62385]: DEBUG oslo_vmware.api [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205937, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.856983] env[62385]: DEBUG oslo_vmware.api [None req-d86e0c75-5a2a-4394-8269-7f8858af71d8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205932, 'name': RemoveSnapshot_Task, 'duration_secs': 1.001003} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.857344] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d86e0c75-5a2a-4394-8269-7f8858af71d8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Deleted Snapshot of the VM instance {{(pid=62385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 863.959949] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63ae7905-9603-4749-b3a8-406618e89cef {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.970840] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-204f8b5f-0ec7-488d-99c2-6c15cf38371e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.974014] env[62385]: DEBUG oslo_vmware.api [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52666ac2-4eff-e013-d15a-dc2b5d6dc2ac, 'name': SearchDatastore_Task, 'duration_secs': 0.011665} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.974682] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.974965] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 863.975258] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.975447] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.975669] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 863.976370] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cdcd893d-a15a-482a-bd06-4039613857f6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.008917] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff91dfe9-b7e7-42d5-8ff6-8869a3a5b2f4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.015582] env[62385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 864.015582] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5267ea8f-cbbc-8b54-2b2b-ac7c53e52d7b" [ 864.015582] env[62385]: _type = "HttpNfcLease" [ 864.015582] env[62385]: } is ready. {{(pid=62385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 864.017126] env[62385]: DEBUG oslo_vmware.rw_handles [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 864.017126] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5267ea8f-cbbc-8b54-2b2b-ac7c53e52d7b" [ 864.017126] env[62385]: _type = "HttpNfcLease" [ 864.017126] env[62385]: }. {{(pid=62385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 864.017459] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 864.017580] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 864.018731] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efefa2da-0e6c-4e9a-a72d-f35abbe5b84d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.021599] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a6430e0c-3205-44af-9767-187ac4735215 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.027037] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c30268e0-a2ff-42fb-9515-9978254771b2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.035319] env[62385]: DEBUG oslo_vmware.api [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 864.035319] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]520bedef-2cd5-c35c-91e8-84fffe868e14" [ 864.035319] env[62385]: _type = "Task" [ 864.035319] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.035565] env[62385]: DEBUG oslo_vmware.rw_handles [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ada463-7601-a7d8-4deb-9352fca7de64/disk-0.vmdk from lease info. {{(pid=62385) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 864.035841] env[62385]: DEBUG oslo_vmware.rw_handles [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ada463-7601-a7d8-4deb-9352fca7de64/disk-0.vmdk for reading. {{(pid=62385) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 864.049319] env[62385]: DEBUG nova.compute.provider_tree [None req-5dfcbc54-c64f-4274-aed3-ea9558593cc5 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 864.112739] env[62385]: DEBUG oslo_vmware.api [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]520bedef-2cd5-c35c-91e8-84fffe868e14, 'name': SearchDatastore_Task, 'duration_secs': 0.064794} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.115204] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d33281f-20ba-4c9b-88eb-b42b4b888da3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.122971] env[62385]: DEBUG oslo_vmware.api [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 864.122971] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a0009c-abb5-f5e2-b719-c62952e526f5" [ 864.122971] env[62385]: _type = "Task" [ 864.122971] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.133065] env[62385]: DEBUG oslo_vmware.api [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a0009c-abb5-f5e2-b719-c62952e526f5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.183823] env[62385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-6945f518-19a0-4b9c-94a9-17af44d4e3bf {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.246882] env[62385]: DEBUG oslo_vmware.api [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205936, 'name': PowerOnVM_Task} progress is 78%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.304063] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquiring lock "refresh_cache-bd7263b2-b996-4794-946b-2c28215574cb" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.304063] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquired lock "refresh_cache-bd7263b2-b996-4794-946b-2c28215574cb" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.304224] env[62385]: DEBUG nova.network.neutron [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 864.322481] env[62385]: DEBUG oslo_vmware.api [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205937, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.366770] env[62385]: WARNING nova.compute.manager [None req-d86e0c75-5a2a-4394-8269-7f8858af71d8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Image not found during snapshot: nova.exception.ImageNotFound: Image e3323aab-9833-47a5-ab60-9095f41291a9 could not be found. [ 864.370642] env[62385]: DEBUG oslo_concurrency.lockutils [None req-032bafda-a74f-4b9e-8041-96a5dea4e54f tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Acquiring lock "699a57a9-5a1a-4cd1-8449-723400d9caf4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.370884] env[62385]: DEBUG oslo_concurrency.lockutils [None req-032bafda-a74f-4b9e-8041-96a5dea4e54f tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Lock "699a57a9-5a1a-4cd1-8449-723400d9caf4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.371137] env[62385]: DEBUG oslo_concurrency.lockutils [None req-032bafda-a74f-4b9e-8041-96a5dea4e54f tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Acquiring lock "699a57a9-5a1a-4cd1-8449-723400d9caf4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.371358] env[62385]: DEBUG oslo_concurrency.lockutils [None req-032bafda-a74f-4b9e-8041-96a5dea4e54f tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Lock "699a57a9-5a1a-4cd1-8449-723400d9caf4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.371546] env[62385]: DEBUG oslo_concurrency.lockutils [None req-032bafda-a74f-4b9e-8041-96a5dea4e54f tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Lock "699a57a9-5a1a-4cd1-8449-723400d9caf4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.374025] env[62385]: INFO nova.compute.manager [None req-032bafda-a74f-4b9e-8041-96a5dea4e54f tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Terminating instance [ 864.377571] env[62385]: DEBUG nova.compute.manager [None req-032bafda-a74f-4b9e-8041-96a5dea4e54f tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 864.377799] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-032bafda-a74f-4b9e-8041-96a5dea4e54f tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 864.379023] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85d42e16-adc8-47ca-8558-5ad155450edc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.389873] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-032bafda-a74f-4b9e-8041-96a5dea4e54f tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 864.390187] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-caa7e03a-b6d8-4cc7-bf44-5ff7e37a24fe {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.398642] env[62385]: DEBUG oslo_vmware.api [None req-032bafda-a74f-4b9e-8041-96a5dea4e54f tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Waiting for the task: (returnval){ [ 864.398642] env[62385]: value = "task-1205938" [ 864.398642] env[62385]: _type = "Task" [ 864.398642] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.409267] env[62385]: DEBUG oslo_vmware.api [None req-032bafda-a74f-4b9e-8041-96a5dea4e54f tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Task: {'id': task-1205938, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.537497] env[62385]: DEBUG nova.compute.manager [req-8fe38ce4-bb4a-429f-8937-d288187742c2 req-bb43299d-11b1-4cad-a78a-1d09e9f9db04 service nova] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Received event network-vif-plugged-6820d526-f9c3-43cc-91db-538e9c3bd249 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 864.537766] env[62385]: DEBUG oslo_concurrency.lockutils [req-8fe38ce4-bb4a-429f-8937-d288187742c2 req-bb43299d-11b1-4cad-a78a-1d09e9f9db04 service nova] Acquiring lock "fbc71ada-d3b5-46f6-90a9-489c118d5126-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.538037] env[62385]: DEBUG oslo_concurrency.lockutils [req-8fe38ce4-bb4a-429f-8937-d288187742c2 req-bb43299d-11b1-4cad-a78a-1d09e9f9db04 service nova] Lock "fbc71ada-d3b5-46f6-90a9-489c118d5126-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.538248] env[62385]: DEBUG oslo_concurrency.lockutils [req-8fe38ce4-bb4a-429f-8937-d288187742c2 req-bb43299d-11b1-4cad-a78a-1d09e9f9db04 service nova] Lock "fbc71ada-d3b5-46f6-90a9-489c118d5126-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.538474] env[62385]: DEBUG nova.compute.manager [req-8fe38ce4-bb4a-429f-8937-d288187742c2 req-bb43299d-11b1-4cad-a78a-1d09e9f9db04 service nova] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] No waiting events found dispatching network-vif-plugged-6820d526-f9c3-43cc-91db-538e9c3bd249 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 864.538674] env[62385]: WARNING nova.compute.manager [req-8fe38ce4-bb4a-429f-8937-d288187742c2 req-bb43299d-11b1-4cad-a78a-1d09e9f9db04 service nova] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Received unexpected event network-vif-plugged-6820d526-f9c3-43cc-91db-538e9c3bd249 for instance with vm_state building and task_state spawning. [ 864.538911] env[62385]: DEBUG nova.compute.manager [req-8fe38ce4-bb4a-429f-8937-d288187742c2 req-bb43299d-11b1-4cad-a78a-1d09e9f9db04 service nova] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Received event network-changed-6820d526-f9c3-43cc-91db-538e9c3bd249 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 864.539150] env[62385]: DEBUG nova.compute.manager [req-8fe38ce4-bb4a-429f-8937-d288187742c2 req-bb43299d-11b1-4cad-a78a-1d09e9f9db04 service nova] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Refreshing instance network info cache due to event network-changed-6820d526-f9c3-43cc-91db-538e9c3bd249. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 864.539356] env[62385]: DEBUG oslo_concurrency.lockutils [req-8fe38ce4-bb4a-429f-8937-d288187742c2 req-bb43299d-11b1-4cad-a78a-1d09e9f9db04 service nova] Acquiring lock "refresh_cache-fbc71ada-d3b5-46f6-90a9-489c118d5126" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.539502] env[62385]: DEBUG oslo_concurrency.lockutils [req-8fe38ce4-bb4a-429f-8937-d288187742c2 req-bb43299d-11b1-4cad-a78a-1d09e9f9db04 service nova] Acquired lock "refresh_cache-fbc71ada-d3b5-46f6-90a9-489c118d5126" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.539688] env[62385]: DEBUG nova.network.neutron [req-8fe38ce4-bb4a-429f-8937-d288187742c2 req-bb43299d-11b1-4cad-a78a-1d09e9f9db04 service nova] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Refreshing network info cache for port 6820d526-f9c3-43cc-91db-538e9c3bd249 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 864.607940] env[62385]: DEBUG nova.scheduler.client.report [None req-5dfcbc54-c64f-4274-aed3-ea9558593cc5 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 864.636253] env[62385]: DEBUG oslo_vmware.api [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a0009c-abb5-f5e2-b719-c62952e526f5, 'name': SearchDatastore_Task, 'duration_secs': 0.013258} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.636650] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.637112] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] fe0624a9-09a8-498b-bb3c-fda6cab92341/fe0624a9-09a8-498b-bb3c-fda6cab92341.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 864.637476] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bbad2c29-ed61-4e58-961a-8e4d43a07869 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.647204] env[62385]: DEBUG oslo_vmware.api [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 864.647204] env[62385]: value = "task-1205939" [ 864.647204] env[62385]: _type = "Task" [ 864.647204] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.657534] env[62385]: DEBUG oslo_vmware.api [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205939, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.747128] env[62385]: DEBUG oslo_vmware.api [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205936, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.820063] env[62385]: DEBUG oslo_vmware.api [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205937, 'name': ReconfigVM_Task, 'duration_secs': 0.796527} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.820477] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Reconfigured VM instance instance-0000004a to attach disk [datastore1] 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c/c0abbb8d-77e7-4f22-a256-0faf0a781109-rescue.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 864.821417] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d27bc720-ae1d-45b4-9ea7-2b4dfcbe872c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.843311] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7b5bd489-2f81-4572-8b68-daa320f73dc8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Acquiring lock "2a579d13-5372-4340-b7b3-cc02c1912624" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.843649] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7b5bd489-2f81-4572-8b68-daa320f73dc8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Lock "2a579d13-5372-4340-b7b3-cc02c1912624" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.843910] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7b5bd489-2f81-4572-8b68-daa320f73dc8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Acquiring lock "2a579d13-5372-4340-b7b3-cc02c1912624-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.844194] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7b5bd489-2f81-4572-8b68-daa320f73dc8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Lock "2a579d13-5372-4340-b7b3-cc02c1912624-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.844479] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7b5bd489-2f81-4572-8b68-daa320f73dc8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Lock "2a579d13-5372-4340-b7b3-cc02c1912624-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.851935] env[62385]: INFO nova.compute.manager [None req-7b5bd489-2f81-4572-8b68-daa320f73dc8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Terminating instance [ 864.853577] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b8ef767b-9522-400c-8773-df39e4241e33 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.864992] env[62385]: DEBUG nova.network.neutron [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 864.867339] env[62385]: DEBUG nova.compute.manager [None req-7b5bd489-2f81-4572-8b68-daa320f73dc8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 864.867657] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-7b5bd489-2f81-4572-8b68-daa320f73dc8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 864.868713] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47190a0f-d00e-4656-b8cf-91df4a18efed {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.877876] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b5bd489-2f81-4572-8b68-daa320f73dc8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 864.879365] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b8ab7dc6-4d11-48be-a77a-2f816a1119f8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.881174] env[62385]: DEBUG oslo_vmware.api [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for the task: (returnval){ [ 864.881174] env[62385]: value = "task-1205940" [ 864.881174] env[62385]: _type = "Task" [ 864.881174] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.886859] env[62385]: DEBUG oslo_vmware.api [None req-7b5bd489-2f81-4572-8b68-daa320f73dc8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for the task: (returnval){ [ 864.886859] env[62385]: value = "task-1205941" [ 864.886859] env[62385]: _type = "Task" [ 864.886859] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.893251] env[62385]: DEBUG oslo_vmware.api [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205940, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.901502] env[62385]: DEBUG oslo_vmware.api [None req-7b5bd489-2f81-4572-8b68-daa320f73dc8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205941, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.913844] env[62385]: DEBUG oslo_vmware.api [None req-032bafda-a74f-4b9e-8041-96a5dea4e54f tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Task: {'id': task-1205938, 'name': PowerOffVM_Task, 'duration_secs': 0.218766} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.914220] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-032bafda-a74f-4b9e-8041-96a5dea4e54f tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 864.914412] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-032bafda-a74f-4b9e-8041-96a5dea4e54f tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 864.914649] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-66bcfc8a-6446-4e41-8b07-caaebc8a5a18 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.111419] env[62385]: DEBUG nova.network.neutron [req-8fe38ce4-bb4a-429f-8937-d288187742c2 req-bb43299d-11b1-4cad-a78a-1d09e9f9db04 service nova] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 865.113925] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5dfcbc54-c64f-4274-aed3-ea9558593cc5 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.104s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.116796] env[62385]: DEBUG oslo_concurrency.lockutils [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.131s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.119853] env[62385]: INFO nova.compute.claims [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 865.144809] env[62385]: INFO nova.scheduler.client.report [None req-5dfcbc54-c64f-4274-aed3-ea9558593cc5 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Deleted allocations for instance 0d823fcc-cbba-41f6-a47c-2f57d3e79948 [ 865.166358] env[62385]: DEBUG oslo_vmware.api [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205939, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.209069] env[62385]: DEBUG nova.network.neutron [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Updating instance_info_cache with network_info: [{"id": "fa9f8b71-802a-46a3-a93b-20ee02c6adcb", "address": "fa:16:3e:8d:a4:23", "network": {"id": "2c1e2031-ab59-47f5-a758-3a314aca6c1e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1100903482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "731b6700ce5b43369b40165545f04950", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1316f5aa-529f-4bac-8dd7-6076a9d43312", "external-id": "nsx-vlan-transportzone-399", "segmentation_id": 399, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa9f8b71-80", "ovs_interfaceid": "fa9f8b71-802a-46a3-a93b-20ee02c6adcb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.253236] env[62385]: DEBUG oslo_vmware.api [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205936, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.264851] env[62385]: DEBUG nova.network.neutron [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Successfully updated port: 2fdc640e-c457-46d4-8a9d-fe1f934bd29d {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 865.272347] env[62385]: DEBUG nova.compute.manager [req-5b60c21c-a279-4a99-a2a7-c4dc28fc45f6 req-b5c6ce36-231d-46a4-9510-8af7727622e3 service nova] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Received event network-vif-plugged-2fdc640e-c457-46d4-8a9d-fe1f934bd29d {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 865.272630] env[62385]: DEBUG oslo_concurrency.lockutils [req-5b60c21c-a279-4a99-a2a7-c4dc28fc45f6 req-b5c6ce36-231d-46a4-9510-8af7727622e3 service nova] Acquiring lock "fbc71ada-d3b5-46f6-90a9-489c118d5126-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.272907] env[62385]: DEBUG oslo_concurrency.lockutils [req-5b60c21c-a279-4a99-a2a7-c4dc28fc45f6 req-b5c6ce36-231d-46a4-9510-8af7727622e3 service nova] Lock "fbc71ada-d3b5-46f6-90a9-489c118d5126-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.273488] env[62385]: DEBUG oslo_concurrency.lockutils [req-5b60c21c-a279-4a99-a2a7-c4dc28fc45f6 req-b5c6ce36-231d-46a4-9510-8af7727622e3 service nova] Lock "fbc71ada-d3b5-46f6-90a9-489c118d5126-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.273758] env[62385]: DEBUG nova.compute.manager [req-5b60c21c-a279-4a99-a2a7-c4dc28fc45f6 req-b5c6ce36-231d-46a4-9510-8af7727622e3 service nova] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] No waiting events found dispatching network-vif-plugged-2fdc640e-c457-46d4-8a9d-fe1f934bd29d {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 865.274018] env[62385]: WARNING nova.compute.manager [req-5b60c21c-a279-4a99-a2a7-c4dc28fc45f6 req-b5c6ce36-231d-46a4-9510-8af7727622e3 service nova] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Received unexpected event network-vif-plugged-2fdc640e-c457-46d4-8a9d-fe1f934bd29d for instance with vm_state building and task_state spawning. [ 865.338593] env[62385]: DEBUG nova.network.neutron [req-8fe38ce4-bb4a-429f-8937-d288187742c2 req-bb43299d-11b1-4cad-a78a-1d09e9f9db04 service nova] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.393235] env[62385]: DEBUG oslo_vmware.api [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205940, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.404271] env[62385]: DEBUG oslo_vmware.api [None req-7b5bd489-2f81-4572-8b68-daa320f73dc8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205941, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.666809] env[62385]: DEBUG oslo_vmware.api [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205939, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.667681] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5dfcbc54-c64f-4274-aed3-ea9558593cc5 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "0d823fcc-cbba-41f6-a47c-2f57d3e79948" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.485s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.712142] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Releasing lock "refresh_cache-bd7263b2-b996-4794-946b-2c28215574cb" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.712427] env[62385]: DEBUG nova.compute.manager [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Instance network_info: |[{"id": "fa9f8b71-802a-46a3-a93b-20ee02c6adcb", "address": "fa:16:3e:8d:a4:23", "network": {"id": "2c1e2031-ab59-47f5-a758-3a314aca6c1e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1100903482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "731b6700ce5b43369b40165545f04950", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1316f5aa-529f-4bac-8dd7-6076a9d43312", "external-id": "nsx-vlan-transportzone-399", "segmentation_id": 399, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa9f8b71-80", "ovs_interfaceid": "fa9f8b71-802a-46a3-a93b-20ee02c6adcb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 865.712871] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8d:a4:23', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1316f5aa-529f-4bac-8dd7-6076a9d43312', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fa9f8b71-802a-46a3-a93b-20ee02c6adcb', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 865.721191] env[62385]: DEBUG oslo.service.loopingcall [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 865.721777] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 865.722723] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3d5f59d0-fa89-40ea-a443-225c298dc9a5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.748838] env[62385]: DEBUG oslo_vmware.api [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205936, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.750511] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 865.750511] env[62385]: value = "task-1205943" [ 865.750511] env[62385]: _type = "Task" [ 865.750511] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.761333] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205943, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.770151] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Acquiring lock "refresh_cache-fbc71ada-d3b5-46f6-90a9-489c118d5126" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.842254] env[62385]: DEBUG oslo_concurrency.lockutils [req-8fe38ce4-bb4a-429f-8937-d288187742c2 req-bb43299d-11b1-4cad-a78a-1d09e9f9db04 service nova] Releasing lock "refresh_cache-fbc71ada-d3b5-46f6-90a9-489c118d5126" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.842754] env[62385]: DEBUG nova.compute.manager [req-8fe38ce4-bb4a-429f-8937-d288187742c2 req-bb43299d-11b1-4cad-a78a-1d09e9f9db04 service nova] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Received event network-vif-plugged-fa9f8b71-802a-46a3-a93b-20ee02c6adcb {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 865.842924] env[62385]: DEBUG oslo_concurrency.lockutils [req-8fe38ce4-bb4a-429f-8937-d288187742c2 req-bb43299d-11b1-4cad-a78a-1d09e9f9db04 service nova] Acquiring lock "bd7263b2-b996-4794-946b-2c28215574cb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.843313] env[62385]: DEBUG oslo_concurrency.lockutils [req-8fe38ce4-bb4a-429f-8937-d288187742c2 req-bb43299d-11b1-4cad-a78a-1d09e9f9db04 service nova] Lock "bd7263b2-b996-4794-946b-2c28215574cb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.843609] env[62385]: DEBUG oslo_concurrency.lockutils [req-8fe38ce4-bb4a-429f-8937-d288187742c2 req-bb43299d-11b1-4cad-a78a-1d09e9f9db04 service nova] Lock "bd7263b2-b996-4794-946b-2c28215574cb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.843901] env[62385]: DEBUG nova.compute.manager [req-8fe38ce4-bb4a-429f-8937-d288187742c2 req-bb43299d-11b1-4cad-a78a-1d09e9f9db04 service nova] [instance: bd7263b2-b996-4794-946b-2c28215574cb] No waiting events found dispatching network-vif-plugged-fa9f8b71-802a-46a3-a93b-20ee02c6adcb {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 865.844227] env[62385]: WARNING nova.compute.manager [req-8fe38ce4-bb4a-429f-8937-d288187742c2 req-bb43299d-11b1-4cad-a78a-1d09e9f9db04 service nova] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Received unexpected event network-vif-plugged-fa9f8b71-802a-46a3-a93b-20ee02c6adcb for instance with vm_state building and task_state spawning. [ 865.844466] env[62385]: DEBUG nova.compute.manager [req-8fe38ce4-bb4a-429f-8937-d288187742c2 req-bb43299d-11b1-4cad-a78a-1d09e9f9db04 service nova] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Received event network-changed-fa9f8b71-802a-46a3-a93b-20ee02c6adcb {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 865.844798] env[62385]: DEBUG nova.compute.manager [req-8fe38ce4-bb4a-429f-8937-d288187742c2 req-bb43299d-11b1-4cad-a78a-1d09e9f9db04 service nova] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Refreshing instance network info cache due to event network-changed-fa9f8b71-802a-46a3-a93b-20ee02c6adcb. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 865.845113] env[62385]: DEBUG oslo_concurrency.lockutils [req-8fe38ce4-bb4a-429f-8937-d288187742c2 req-bb43299d-11b1-4cad-a78a-1d09e9f9db04 service nova] Acquiring lock "refresh_cache-bd7263b2-b996-4794-946b-2c28215574cb" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.845303] env[62385]: DEBUG oslo_concurrency.lockutils [req-8fe38ce4-bb4a-429f-8937-d288187742c2 req-bb43299d-11b1-4cad-a78a-1d09e9f9db04 service nova] Acquired lock "refresh_cache-bd7263b2-b996-4794-946b-2c28215574cb" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.845536] env[62385]: DEBUG nova.network.neutron [req-8fe38ce4-bb4a-429f-8937-d288187742c2 req-bb43299d-11b1-4cad-a78a-1d09e9f9db04 service nova] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Refreshing network info cache for port fa9f8b71-802a-46a3-a93b-20ee02c6adcb {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 865.847065] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Acquired lock "refresh_cache-fbc71ada-d3b5-46f6-90a9-489c118d5126" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.847326] env[62385]: DEBUG nova.network.neutron [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 865.894672] env[62385]: DEBUG oslo_vmware.api [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205940, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.905384] env[62385]: DEBUG oslo_vmware.api [None req-7b5bd489-2f81-4572-8b68-daa320f73dc8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205941, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.162041] env[62385]: DEBUG oslo_vmware.api [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205939, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.472069} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.162366] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] fe0624a9-09a8-498b-bb3c-fda6cab92341/fe0624a9-09a8-498b-bb3c-fda6cab92341.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 866.162667] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 866.163050] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3ff7ff5b-3b3e-4bab-86f5-ec7163ba07d2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.173873] env[62385]: DEBUG oslo_vmware.api [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 866.173873] env[62385]: value = "task-1205944" [ 866.173873] env[62385]: _type = "Task" [ 866.173873] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.184943] env[62385]: DEBUG oslo_vmware.api [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205944, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.248993] env[62385]: DEBUG oslo_vmware.api [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205936, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.260637] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205943, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.397897] env[62385]: DEBUG oslo_vmware.api [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205940, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.405336] env[62385]: DEBUG nova.network.neutron [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 866.415902] env[62385]: DEBUG oslo_vmware.api [None req-7b5bd489-2f81-4572-8b68-daa320f73dc8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205941, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.504029] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae8ae34c-3ec2-4407-9be7-edd2af16e38b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.513223] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d42dfad0-be56-487c-a335-2b3f50a71d36 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.549367] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f24f175a-f9eb-4aa5-ac72-d028b5e3aa23 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.560475] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cfe4db0-cf71-422d-84ff-a47f577a3bef {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.576061] env[62385]: DEBUG nova.compute.provider_tree [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 866.689038] env[62385]: DEBUG oslo_vmware.api [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205944, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.110534} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.689038] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 866.689038] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-793f265b-7b79-4587-974f-3f0f05225f97 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.713989] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] fe0624a9-09a8-498b-bb3c-fda6cab92341/fe0624a9-09a8-498b-bb3c-fda6cab92341.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 866.714640] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cbc35a52-f77b-4fbf-b6df-d6eab6f801a6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.743961] env[62385]: DEBUG oslo_vmware.api [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 866.743961] env[62385]: value = "task-1205945" [ 866.743961] env[62385]: _type = "Task" [ 866.743961] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.753977] env[62385]: DEBUG oslo_vmware.api [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205936, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.761105] env[62385]: DEBUG oslo_vmware.api [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205945, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.766853] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205943, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.894292] env[62385]: DEBUG oslo_vmware.api [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205940, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.895455] env[62385]: DEBUG nova.network.neutron [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Updating instance_info_cache with network_info: [{"id": "2572a945-d7b9-47e5-8641-3718434e80ca", "address": "fa:16:3e:80:c2:26", "network": {"id": "0e19be42-16f3-43a9-9921-1f763804af82", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-846114175", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.182", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d68661a0197d4e3cacc06ea433ac706e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7edb7c08-2fae-4df5-9ec6-5ccf06d7e337", "external-id": "nsx-vlan-transportzone-309", "segmentation_id": 309, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2572a945-d7", "ovs_interfaceid": "2572a945-d7b9-47e5-8641-3718434e80ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "6820d526-f9c3-43cc-91db-538e9c3bd249", "address": "fa:16:3e:1a:d1:98", "network": {"id": "24d1af4a-afeb-449c-9ff6-367ea00a6763", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-990399095", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.204", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "d68661a0197d4e3cacc06ea433ac706e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c02dd284-ab80-451c-93eb-48c8360acb9c", "external-id": "nsx-vlan-transportzone-818", "segmentation_id": 818, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6820d526-f9", "ovs_interfaceid": "6820d526-f9c3-43cc-91db-538e9c3bd249", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "2fdc640e-c457-46d4-8a9d-fe1f934bd29d", "address": "fa:16:3e:1a:b8:3b", "network": {"id": "0e19be42-16f3-43a9-9921-1f763804af82", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-846114175", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.34", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d68661a0197d4e3cacc06ea433ac706e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7edb7c08-2fae-4df5-9ec6-5ccf06d7e337", "external-id": "nsx-vlan-transportzone-309", "segmentation_id": 309, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2fdc640e-c4", "ovs_interfaceid": "2fdc640e-c457-46d4-8a9d-fe1f934bd29d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.913340] env[62385]: DEBUG oslo_vmware.api [None req-7b5bd489-2f81-4572-8b68-daa320f73dc8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205941, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.079716] env[62385]: DEBUG nova.scheduler.client.report [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 867.212028] env[62385]: DEBUG nova.network.neutron [req-8fe38ce4-bb4a-429f-8937-d288187742c2 req-bb43299d-11b1-4cad-a78a-1d09e9f9db04 service nova] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Updated VIF entry in instance network info cache for port fa9f8b71-802a-46a3-a93b-20ee02c6adcb. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 867.212585] env[62385]: DEBUG nova.network.neutron [req-8fe38ce4-bb4a-429f-8937-d288187742c2 req-bb43299d-11b1-4cad-a78a-1d09e9f9db04 service nova] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Updating instance_info_cache with network_info: [{"id": "fa9f8b71-802a-46a3-a93b-20ee02c6adcb", "address": "fa:16:3e:8d:a4:23", "network": {"id": "2c1e2031-ab59-47f5-a758-3a314aca6c1e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1100903482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "731b6700ce5b43369b40165545f04950", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1316f5aa-529f-4bac-8dd7-6076a9d43312", "external-id": "nsx-vlan-transportzone-399", "segmentation_id": 399, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa9f8b71-80", "ovs_interfaceid": "fa9f8b71-802a-46a3-a93b-20ee02c6adcb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.251208] env[62385]: DEBUG oslo_vmware.api [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205936, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.266118] env[62385]: DEBUG oslo_vmware.api [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205945, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.271118] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205943, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.301102] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-032bafda-a74f-4b9e-8041-96a5dea4e54f tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 867.301452] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-032bafda-a74f-4b9e-8041-96a5dea4e54f tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 867.301663] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-032bafda-a74f-4b9e-8041-96a5dea4e54f tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Deleting the datastore file [datastore1] 699a57a9-5a1a-4cd1-8449-723400d9caf4 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 867.303112] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9ac6f9bc-da63-495a-bff6-c6bc093b7177 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.331052] env[62385]: DEBUG oslo_vmware.api [None req-032bafda-a74f-4b9e-8041-96a5dea4e54f tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Waiting for the task: (returnval){ [ 867.331052] env[62385]: value = "task-1205946" [ 867.331052] env[62385]: _type = "Task" [ 867.331052] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.344591] env[62385]: DEBUG oslo_vmware.api [None req-032bafda-a74f-4b9e-8041-96a5dea4e54f tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Task: {'id': task-1205946, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.395405] env[62385]: DEBUG oslo_vmware.api [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205940, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.398268] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Releasing lock "refresh_cache-fbc71ada-d3b5-46f6-90a9-489c118d5126" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.398913] env[62385]: DEBUG nova.compute.manager [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Instance network_info: |[{"id": "2572a945-d7b9-47e5-8641-3718434e80ca", "address": "fa:16:3e:80:c2:26", "network": {"id": "0e19be42-16f3-43a9-9921-1f763804af82", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-846114175", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.182", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d68661a0197d4e3cacc06ea433ac706e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7edb7c08-2fae-4df5-9ec6-5ccf06d7e337", "external-id": "nsx-vlan-transportzone-309", "segmentation_id": 309, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2572a945-d7", "ovs_interfaceid": "2572a945-d7b9-47e5-8641-3718434e80ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "6820d526-f9c3-43cc-91db-538e9c3bd249", "address": "fa:16:3e:1a:d1:98", "network": {"id": "24d1af4a-afeb-449c-9ff6-367ea00a6763", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-990399095", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.204", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "d68661a0197d4e3cacc06ea433ac706e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c02dd284-ab80-451c-93eb-48c8360acb9c", "external-id": "nsx-vlan-transportzone-818", "segmentation_id": 818, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6820d526-f9", "ovs_interfaceid": "6820d526-f9c3-43cc-91db-538e9c3bd249", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "2fdc640e-c457-46d4-8a9d-fe1f934bd29d", "address": "fa:16:3e:1a:b8:3b", "network": {"id": "0e19be42-16f3-43a9-9921-1f763804af82", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-846114175", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.34", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d68661a0197d4e3cacc06ea433ac706e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7edb7c08-2fae-4df5-9ec6-5ccf06d7e337", "external-id": "nsx-vlan-transportzone-309", "segmentation_id": 309, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2fdc640e-c4", "ovs_interfaceid": "2fdc640e-c457-46d4-8a9d-fe1f934bd29d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 867.399685] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:80:c2:26', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7edb7c08-2fae-4df5-9ec6-5ccf06d7e337', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2572a945-d7b9-47e5-8641-3718434e80ca', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:1a:d1:98', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c02dd284-ab80-451c-93eb-48c8360acb9c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6820d526-f9c3-43cc-91db-538e9c3bd249', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:1a:b8:3b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7edb7c08-2fae-4df5-9ec6-5ccf06d7e337', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2fdc640e-c457-46d4-8a9d-fe1f934bd29d', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 867.412637] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Creating folder: Project (d68661a0197d4e3cacc06ea433ac706e). Parent ref: group-v261107. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 867.413220] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-985d7e92-e6e1-4905-873a-510f518cbb5d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.425145] env[62385]: DEBUG oslo_vmware.api [None req-7b5bd489-2f81-4572-8b68-daa320f73dc8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205941, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.431357] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Created folder: Project (d68661a0197d4e3cacc06ea433ac706e) in parent group-v261107. [ 867.431863] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Creating folder: Instances. Parent ref: group-v261233. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 867.432295] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6273b02c-9e9b-4a82-b045-8a0619ab8939 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.446914] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Created folder: Instances in parent group-v261233. [ 867.446914] env[62385]: DEBUG oslo.service.loopingcall [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 867.446914] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 867.446914] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f094e376-5a84-4757-b0e7-96498bb0ba63 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.484211] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 867.484211] env[62385]: value = "task-1205949" [ 867.484211] env[62385]: _type = "Task" [ 867.484211] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.495218] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205949, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.503285] env[62385]: DEBUG nova.compute.manager [req-5575d1e3-e0af-45eb-805a-fca8694c48e7 req-62bfa948-7e44-43ee-a166-350cbc9bfe5c service nova] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Received event network-changed-2fdc640e-c457-46d4-8a9d-fe1f934bd29d {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 867.503285] env[62385]: DEBUG nova.compute.manager [req-5575d1e3-e0af-45eb-805a-fca8694c48e7 req-62bfa948-7e44-43ee-a166-350cbc9bfe5c service nova] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Refreshing instance network info cache due to event network-changed-2fdc640e-c457-46d4-8a9d-fe1f934bd29d. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 867.503285] env[62385]: DEBUG oslo_concurrency.lockutils [req-5575d1e3-e0af-45eb-805a-fca8694c48e7 req-62bfa948-7e44-43ee-a166-350cbc9bfe5c service nova] Acquiring lock "refresh_cache-fbc71ada-d3b5-46f6-90a9-489c118d5126" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.503285] env[62385]: DEBUG oslo_concurrency.lockutils [req-5575d1e3-e0af-45eb-805a-fca8694c48e7 req-62bfa948-7e44-43ee-a166-350cbc9bfe5c service nova] Acquired lock "refresh_cache-fbc71ada-d3b5-46f6-90a9-489c118d5126" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.503927] env[62385]: DEBUG nova.network.neutron [req-5575d1e3-e0af-45eb-805a-fca8694c48e7 req-62bfa948-7e44-43ee-a166-350cbc9bfe5c service nova] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Refreshing network info cache for port 2fdc640e-c457-46d4-8a9d-fe1f934bd29d {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 867.587591] env[62385]: DEBUG oslo_concurrency.lockutils [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.469s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.587591] env[62385]: DEBUG nova.compute.manager [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 867.590741] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 16.861s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.590916] env[62385]: DEBUG nova.objects.instance [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62385) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 867.716453] env[62385]: DEBUG oslo_concurrency.lockutils [req-8fe38ce4-bb4a-429f-8937-d288187742c2 req-bb43299d-11b1-4cad-a78a-1d09e9f9db04 service nova] Releasing lock "refresh_cache-bd7263b2-b996-4794-946b-2c28215574cb" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.753576] env[62385]: DEBUG oslo_vmware.api [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205936, 'name': PowerOnVM_Task, 'duration_secs': 3.576321} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.756874] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 867.757105] env[62385]: INFO nova.compute.manager [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Took 13.75 seconds to spawn the instance on the hypervisor. [ 867.757298] env[62385]: DEBUG nova.compute.manager [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 867.758436] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17ea256c-49fb-443d-a267-32f6225c1115 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.780845] env[62385]: DEBUG oslo_vmware.api [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205945, 'name': ReconfigVM_Task, 'duration_secs': 0.584901} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.784504] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205943, 'name': CreateVM_Task, 'duration_secs': 1.99277} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.785069] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Reconfigured VM instance instance-0000004e to attach disk [datastore2] fe0624a9-09a8-498b-bb3c-fda6cab92341/fe0624a9-09a8-498b-bb3c-fda6cab92341.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 867.785669] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 867.786144] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-344d3020-ae18-407e-8048-8027c1902be0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.788711] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.788949] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.789318] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 867.789960] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-44bd9709-1334-46b4-85db-d70c2b0016b2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.795983] env[62385]: DEBUG oslo_vmware.api [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 867.795983] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52f4aaba-70fc-4e07-3999-599d18458ea5" [ 867.795983] env[62385]: _type = "Task" [ 867.795983] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.800588] env[62385]: DEBUG oslo_vmware.api [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 867.800588] env[62385]: value = "task-1205950" [ 867.800588] env[62385]: _type = "Task" [ 867.800588] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.808432] env[62385]: DEBUG oslo_vmware.api [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52f4aaba-70fc-4e07-3999-599d18458ea5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.815227] env[62385]: DEBUG oslo_vmware.api [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205950, 'name': Rename_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.841410] env[62385]: DEBUG oslo_vmware.api [None req-032bafda-a74f-4b9e-8041-96a5dea4e54f tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Task: {'id': task-1205946, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.26369} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.841706] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-032bafda-a74f-4b9e-8041-96a5dea4e54f tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 867.841917] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-032bafda-a74f-4b9e-8041-96a5dea4e54f tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 867.842161] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-032bafda-a74f-4b9e-8041-96a5dea4e54f tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 867.842295] env[62385]: INFO nova.compute.manager [None req-032bafda-a74f-4b9e-8041-96a5dea4e54f tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Took 3.46 seconds to destroy the instance on the hypervisor. [ 867.842562] env[62385]: DEBUG oslo.service.loopingcall [None req-032bafda-a74f-4b9e-8041-96a5dea4e54f tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 867.842769] env[62385]: DEBUG nova.compute.manager [-] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 867.842868] env[62385]: DEBUG nova.network.neutron [-] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 867.896329] env[62385]: DEBUG oslo_vmware.api [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205940, 'name': ReconfigVM_Task, 'duration_secs': 2.673139} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.896789] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 867.896907] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1c1eb3d0-f8f4-407d-8ca7-04feb15b9408 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.903994] env[62385]: DEBUG oslo_vmware.api [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for the task: (returnval){ [ 867.903994] env[62385]: value = "task-1205951" [ 867.903994] env[62385]: _type = "Task" [ 867.903994] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.912743] env[62385]: DEBUG oslo_vmware.api [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205951, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.922193] env[62385]: DEBUG oslo_vmware.api [None req-7b5bd489-2f81-4572-8b68-daa320f73dc8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205941, 'name': PowerOffVM_Task, 'duration_secs': 2.656379} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.922475] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b5bd489-2f81-4572-8b68-daa320f73dc8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 867.922737] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-7b5bd489-2f81-4572-8b68-daa320f73dc8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 867.923033] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0facfd05-daea-4c9e-9595-badba76f3ab4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.996582] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205949, 'name': CreateVM_Task} progress is 25%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.010963] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-7b5bd489-2f81-4572-8b68-daa320f73dc8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 868.011276] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-7b5bd489-2f81-4572-8b68-daa320f73dc8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 868.011510] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b5bd489-2f81-4572-8b68-daa320f73dc8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Deleting the datastore file [datastore1] 2a579d13-5372-4340-b7b3-cc02c1912624 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 868.011798] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2d3010c5-ae7b-4cbf-9727-271ccaf55015 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.020080] env[62385]: DEBUG oslo_vmware.api [None req-7b5bd489-2f81-4572-8b68-daa320f73dc8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for the task: (returnval){ [ 868.020080] env[62385]: value = "task-1205953" [ 868.020080] env[62385]: _type = "Task" [ 868.020080] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.030146] env[62385]: DEBUG oslo_vmware.api [None req-7b5bd489-2f81-4572-8b68-daa320f73dc8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205953, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.105339] env[62385]: DEBUG nova.compute.utils [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 868.112022] env[62385]: DEBUG nova.compute.manager [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 868.112022] env[62385]: DEBUG nova.network.neutron [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 868.211577] env[62385]: DEBUG nova.policy [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b314c392b9ca405480c27c25364fda81', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '93111f5ba17c47de9691f90b99e03aa8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 868.284537] env[62385]: DEBUG nova.network.neutron [req-5575d1e3-e0af-45eb-805a-fca8694c48e7 req-62bfa948-7e44-43ee-a166-350cbc9bfe5c service nova] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Updated VIF entry in instance network info cache for port 2fdc640e-c457-46d4-8a9d-fe1f934bd29d. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 868.284815] env[62385]: DEBUG nova.network.neutron [req-5575d1e3-e0af-45eb-805a-fca8694c48e7 req-62bfa948-7e44-43ee-a166-350cbc9bfe5c service nova] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Updating instance_info_cache with network_info: [{"id": "2572a945-d7b9-47e5-8641-3718434e80ca", "address": "fa:16:3e:80:c2:26", "network": {"id": "0e19be42-16f3-43a9-9921-1f763804af82", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-846114175", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.182", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d68661a0197d4e3cacc06ea433ac706e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7edb7c08-2fae-4df5-9ec6-5ccf06d7e337", "external-id": "nsx-vlan-transportzone-309", "segmentation_id": 309, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2572a945-d7", "ovs_interfaceid": "2572a945-d7b9-47e5-8641-3718434e80ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "6820d526-f9c3-43cc-91db-538e9c3bd249", "address": "fa:16:3e:1a:d1:98", "network": {"id": "24d1af4a-afeb-449c-9ff6-367ea00a6763", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-990399095", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.204", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "d68661a0197d4e3cacc06ea433ac706e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c02dd284-ab80-451c-93eb-48c8360acb9c", "external-id": "nsx-vlan-transportzone-818", "segmentation_id": 818, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6820d526-f9", "ovs_interfaceid": "6820d526-f9c3-43cc-91db-538e9c3bd249", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "2fdc640e-c457-46d4-8a9d-fe1f934bd29d", "address": "fa:16:3e:1a:b8:3b", "network": {"id": "0e19be42-16f3-43a9-9921-1f763804af82", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-846114175", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.34", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d68661a0197d4e3cacc06ea433ac706e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7edb7c08-2fae-4df5-9ec6-5ccf06d7e337", "external-id": "nsx-vlan-transportzone-309", "segmentation_id": 309, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2fdc640e-c4", "ovs_interfaceid": "2fdc640e-c457-46d4-8a9d-fe1f934bd29d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.298175] env[62385]: INFO nova.compute.manager [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Took 38.02 seconds to build instance. [ 868.315231] env[62385]: DEBUG oslo_vmware.api [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52f4aaba-70fc-4e07-3999-599d18458ea5, 'name': SearchDatastore_Task, 'duration_secs': 0.018716} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.318595] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.318900] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 868.319167] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.319322] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.319668] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 868.320161] env[62385]: DEBUG oslo_vmware.api [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205950, 'name': Rename_Task, 'duration_secs': 0.177519} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.320446] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a78fa97a-b42a-42b2-9420-1c89dab059af {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.322466] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 868.323044] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6b8db0a8-8143-4cf4-a636-8370c9b2d6fa {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.330485] env[62385]: DEBUG oslo_vmware.api [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 868.330485] env[62385]: value = "task-1205954" [ 868.330485] env[62385]: _type = "Task" [ 868.330485] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.334326] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 868.334501] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 868.335517] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06b668a1-bcc7-409e-ba94-27acafcb63fb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.341083] env[62385]: DEBUG oslo_vmware.api [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205954, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.344609] env[62385]: DEBUG oslo_vmware.api [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 868.344609] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ba046b-897e-7b79-4145-5879a7af690f" [ 868.344609] env[62385]: _type = "Task" [ 868.344609] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.353402] env[62385]: DEBUG oslo_vmware.api [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ba046b-897e-7b79-4145-5879a7af690f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.415604] env[62385]: DEBUG oslo_vmware.api [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205951, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.494911] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205949, 'name': CreateVM_Task, 'duration_secs': 0.673872} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.495187] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 868.496041] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.496295] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.496677] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 868.497015] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e09a554e-a463-46cc-b72d-06540751b436 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.503991] env[62385]: DEBUG oslo_vmware.api [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Waiting for the task: (returnval){ [ 868.503991] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]522aee1f-5659-e87e-658f-027550ae474f" [ 868.503991] env[62385]: _type = "Task" [ 868.503991] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.513167] env[62385]: DEBUG oslo_vmware.api [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]522aee1f-5659-e87e-658f-027550ae474f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.530551] env[62385]: DEBUG oslo_vmware.api [None req-7b5bd489-2f81-4572-8b68-daa320f73dc8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1205953, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.29349} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.530842] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b5bd489-2f81-4572-8b68-daa320f73dc8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 868.532778] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-7b5bd489-2f81-4572-8b68-daa320f73dc8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 868.533053] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-7b5bd489-2f81-4572-8b68-daa320f73dc8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 868.533366] env[62385]: INFO nova.compute.manager [None req-7b5bd489-2f81-4572-8b68-daa320f73dc8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Took 3.67 seconds to destroy the instance on the hypervisor. [ 868.533512] env[62385]: DEBUG oslo.service.loopingcall [None req-7b5bd489-2f81-4572-8b68-daa320f73dc8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 868.533728] env[62385]: DEBUG nova.compute.manager [-] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 868.533812] env[62385]: DEBUG nova.network.neutron [-] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 868.613769] env[62385]: DEBUG nova.compute.manager [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 868.621432] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5ddde350-7e8b-4c47-bdbf-d5d5917f7b6e tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.031s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.625029] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.642s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.625029] env[62385]: INFO nova.compute.claims [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 868.791300] env[62385]: DEBUG oslo_concurrency.lockutils [req-5575d1e3-e0af-45eb-805a-fca8694c48e7 req-62bfa948-7e44-43ee-a166-350cbc9bfe5c service nova] Releasing lock "refresh_cache-fbc71ada-d3b5-46f6-90a9-489c118d5126" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.803770] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67742afe-c79f-482c-8ec2-7d7ad1e285db tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "b7a286a2-9a19-4ef1-8a46-6e2a04e22413" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.540s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.842188] env[62385]: DEBUG oslo_vmware.api [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205954, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.855342] env[62385]: DEBUG oslo_vmware.api [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ba046b-897e-7b79-4145-5879a7af690f, 'name': SearchDatastore_Task, 'duration_secs': 0.018782} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.856204] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6680cce0-9e8b-40de-b414-7306fe8a45f9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.863021] env[62385]: DEBUG oslo_vmware.api [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 868.863021] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52fbeeca-9c61-33d2-146b-366bebbc4a94" [ 868.863021] env[62385]: _type = "Task" [ 868.863021] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.872947] env[62385]: DEBUG oslo_vmware.api [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52fbeeca-9c61-33d2-146b-366bebbc4a94, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.914877] env[62385]: DEBUG oslo_vmware.api [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205951, 'name': PowerOnVM_Task, 'duration_secs': 0.58525} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.915193] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 868.921959] env[62385]: DEBUG nova.compute.manager [None req-c08fce70-effb-46d8-b00b-f1d95b49ccf5 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 868.922918] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30042eef-15f1-4cdb-8dc3-10499b3a7884 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.972260] env[62385]: DEBUG nova.network.neutron [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Successfully created port: 0e719241-7a39-42dd-b37e-31cc75a4da74 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 869.018752] env[62385]: DEBUG oslo_vmware.api [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]522aee1f-5659-e87e-658f-027550ae474f, 'name': SearchDatastore_Task, 'duration_secs': 0.016524} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.020129] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.020473] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 869.020772] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.021019] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.021284] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 869.021922] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1f6fe2d4-7339-4203-a707-b60b14c9ea65 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.036186] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 869.036186] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 869.040375] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-370f7cca-6d9a-43c8-a83e-3b467d2ecab8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.049032] env[62385]: DEBUG oslo_vmware.api [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Waiting for the task: (returnval){ [ 869.049032] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521804b2-6246-ed65-da9c-45164fe093d0" [ 869.049032] env[62385]: _type = "Task" [ 869.049032] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.058644] env[62385]: DEBUG oslo_vmware.api [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521804b2-6246-ed65-da9c-45164fe093d0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.123754] env[62385]: DEBUG nova.network.neutron [-] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.135506] env[62385]: DEBUG nova.compute.manager [req-bea016f3-f18a-4a1a-b75e-47616a8c8816 req-c0a28105-7db0-45ee-84bc-9eb1dba7386e service nova] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Received event network-vif-deleted-73603eea-2766-4ac4-a133-015773484ffc {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 869.135823] env[62385]: INFO nova.compute.manager [req-bea016f3-f18a-4a1a-b75e-47616a8c8816 req-c0a28105-7db0-45ee-84bc-9eb1dba7386e service nova] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Neutron deleted interface 73603eea-2766-4ac4-a133-015773484ffc; detaching it from the instance and deleting it from the info cache [ 869.139025] env[62385]: DEBUG nova.network.neutron [req-bea016f3-f18a-4a1a-b75e-47616a8c8816 req-c0a28105-7db0-45ee-84bc-9eb1dba7386e service nova] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.345739] env[62385]: DEBUG oslo_vmware.api [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205954, 'name': PowerOnVM_Task, 'duration_secs': 0.548985} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.346135] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 869.346455] env[62385]: INFO nova.compute.manager [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Took 9.92 seconds to spawn the instance on the hypervisor. [ 869.346707] env[62385]: DEBUG nova.compute.manager [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 869.347609] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee7e26df-77fd-492e-a4d7-3e1950fea41e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.377021] env[62385]: DEBUG oslo_vmware.api [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52fbeeca-9c61-33d2-146b-366bebbc4a94, 'name': SearchDatastore_Task, 'duration_secs': 0.015319} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.377021] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.377021] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] bd7263b2-b996-4794-946b-2c28215574cb/bd7263b2-b996-4794-946b-2c28215574cb.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 869.377021] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-94865e54-beb8-4b9c-9c5b-674ab74885c4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.389742] env[62385]: DEBUG oslo_vmware.api [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 869.389742] env[62385]: value = "task-1205955" [ 869.389742] env[62385]: _type = "Task" [ 869.389742] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.404503] env[62385]: DEBUG oslo_vmware.api [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205955, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.512175] env[62385]: DEBUG nova.network.neutron [-] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.563951] env[62385]: DEBUG oslo_vmware.api [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521804b2-6246-ed65-da9c-45164fe093d0, 'name': SearchDatastore_Task, 'duration_secs': 0.017122} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.564884] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f5a9a489-d710-4b46-9558-ebe781c7c1b5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.571864] env[62385]: DEBUG oslo_vmware.api [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Waiting for the task: (returnval){ [ 869.571864] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52953a1f-a820-ee3e-dc97-3f1f4b2011f0" [ 869.571864] env[62385]: _type = "Task" [ 869.571864] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.582277] env[62385]: DEBUG oslo_vmware.api [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52953a1f-a820-ee3e-dc97-3f1f4b2011f0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.630020] env[62385]: INFO nova.compute.manager [-] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Took 1.79 seconds to deallocate network for instance. [ 869.636032] env[62385]: DEBUG nova.compute.manager [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 869.644439] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c8d3f321-b8ae-4876-977a-113822816034 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.646750] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "f937ef7b-ab39-4d8a-9577-7faabc652aa5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.647315] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "f937ef7b-ab39-4d8a-9577-7faabc652aa5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.658634] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56b14e52-b306-4c17-b0c4-6ccffb5d0d50 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.683884] env[62385]: DEBUG nova.virt.hardware [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 869.684107] env[62385]: DEBUG nova.virt.hardware [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 869.684107] env[62385]: DEBUG nova.virt.hardware [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 869.684320] env[62385]: DEBUG nova.virt.hardware [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 869.684449] env[62385]: DEBUG nova.virt.hardware [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 869.684601] env[62385]: DEBUG nova.virt.hardware [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 869.684920] env[62385]: DEBUG nova.virt.hardware [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 869.685153] env[62385]: DEBUG nova.virt.hardware [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 869.685405] env[62385]: DEBUG nova.virt.hardware [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 869.685638] env[62385]: DEBUG nova.virt.hardware [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 869.685829] env[62385]: DEBUG nova.virt.hardware [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 869.687039] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b089aa93-c345-4938-8143-629098989511 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.704904] env[62385]: DEBUG nova.compute.manager [req-bea016f3-f18a-4a1a-b75e-47616a8c8816 req-c0a28105-7db0-45ee-84bc-9eb1dba7386e service nova] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Detach interface failed, port_id=73603eea-2766-4ac4-a133-015773484ffc, reason: Instance 699a57a9-5a1a-4cd1-8449-723400d9caf4 could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 869.724271] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92eda2be-70dd-428e-881d-0ae895a327e4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.804265] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b5064658-1041-4627-8e74-39b389b439bf tempest-ServersAdminTestJSON-493781504 tempest-ServersAdminTestJSON-493781504-project-admin] Acquiring lock "refresh_cache-b7a286a2-9a19-4ef1-8a46-6e2a04e22413" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.804484] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b5064658-1041-4627-8e74-39b389b439bf tempest-ServersAdminTestJSON-493781504 tempest-ServersAdminTestJSON-493781504-project-admin] Acquired lock "refresh_cache-b7a286a2-9a19-4ef1-8a46-6e2a04e22413" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.804684] env[62385]: DEBUG nova.network.neutron [None req-b5064658-1041-4627-8e74-39b389b439bf tempest-ServersAdminTestJSON-493781504 tempest-ServersAdminTestJSON-493781504-project-admin] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 869.810238] env[62385]: DEBUG nova.compute.manager [req-cdf786de-59ad-41b4-b63a-70011b0db239 req-d604dfe0-7e63-4437-8440-a4964de8637b service nova] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Received event network-vif-deleted-0be859b7-c0bc-435a-908c-49582f2d31da {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 869.868930] env[62385]: INFO nova.compute.manager [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Took 37.31 seconds to build instance. [ 869.899847] env[62385]: DEBUG oslo_vmware.api [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205955, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.017445] env[62385]: INFO nova.compute.manager [-] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Took 1.48 seconds to deallocate network for instance. [ 870.071453] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14af8690-0bfe-46fd-bb81-56f0bfd2cef6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.091359] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-881d4530-1f36-4876-937f-5c7652918bc5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.095801] env[62385]: DEBUG oslo_vmware.api [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52953a1f-a820-ee3e-dc97-3f1f4b2011f0, 'name': SearchDatastore_Task, 'duration_secs': 0.019462} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.096163] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.096452] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] fbc71ada-d3b5-46f6-90a9-489c118d5126/fbc71ada-d3b5-46f6-90a9-489c118d5126.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 870.097320] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b9fc82f4-d6ba-4eb5-bd9c-e84bb3f806bb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.132078] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6444a374-abef-495d-8d66-9a93827c2f8b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.135280] env[62385]: DEBUG oslo_vmware.api [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Waiting for the task: (returnval){ [ 870.135280] env[62385]: value = "task-1205956" [ 870.135280] env[62385]: _type = "Task" [ 870.135280] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.143342] env[62385]: DEBUG oslo_concurrency.lockutils [None req-032bafda-a74f-4b9e-8041-96a5dea4e54f tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.145478] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb23d951-ffc1-40df-90a5-4dfd5cc027a5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.154218] env[62385]: DEBUG nova.compute.manager [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 870.159859] env[62385]: DEBUG oslo_vmware.api [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': task-1205956, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.171629] env[62385]: DEBUG nova.compute.provider_tree [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 870.371656] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ddcea00-4b3e-4f16-a0c2-85e436087fb2 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "fe0624a9-09a8-498b-bb3c-fda6cab92341" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.820s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.402253] env[62385]: DEBUG oslo_vmware.api [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205955, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.769534} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.402570] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] bd7263b2-b996-4794-946b-2c28215574cb/bd7263b2-b996-4794-946b-2c28215574cb.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 870.402820] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 870.403135] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-78ec4c9a-8f70-4ebe-842a-e646ed12ad7d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.414546] env[62385]: DEBUG oslo_vmware.api [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 870.414546] env[62385]: value = "task-1205957" [ 870.414546] env[62385]: _type = "Task" [ 870.414546] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.429455] env[62385]: DEBUG oslo_vmware.api [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205957, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.516457] env[62385]: INFO nova.compute.manager [None req-7766957d-6baa-406f-b861-20b7a4a54b1f tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Unrescuing [ 870.516832] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7766957d-6baa-406f-b861-20b7a4a54b1f tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Acquiring lock "refresh_cache-5e5f43e8-ce45-45c7-a244-bfa724ea5e3c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.517036] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7766957d-6baa-406f-b861-20b7a4a54b1f tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Acquired lock "refresh_cache-5e5f43e8-ce45-45c7-a244-bfa724ea5e3c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.517220] env[62385]: DEBUG nova.network.neutron [None req-7766957d-6baa-406f-b861-20b7a4a54b1f tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 870.530355] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7b5bd489-2f81-4572-8b68-daa320f73dc8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.565765] env[62385]: DEBUG nova.network.neutron [None req-b5064658-1041-4627-8e74-39b389b439bf tempest-ServersAdminTestJSON-493781504 tempest-ServersAdminTestJSON-493781504-project-admin] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Updating instance_info_cache with network_info: [{"id": "2e49e027-be2a-4225-9a34-1e2f9cfb6022", "address": "fa:16:3e:cf:bd:fd", "network": {"id": "b0a3dea4-1925-4920-8a8e-9dc66733d1da", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1951021425-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ac0c8638f064daba042ded4de8c1cc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee20e439-fed9-490e-97dd-f3c886977ae1", "external-id": "nsx-vlan-transportzone-357", "segmentation_id": 357, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e49e027-be", "ovs_interfaceid": "2e49e027-be2a-4225-9a34-1e2f9cfb6022", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.648762] env[62385]: DEBUG oslo_vmware.api [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': task-1205956, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.674674] env[62385]: DEBUG nova.scheduler.client.report [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 870.683620] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.926965] env[62385]: DEBUG oslo_vmware.api [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205957, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.136468} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.926965] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 870.926965] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc26e17c-5571-4d3d-9d98-f6d92ca64bc5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.956980] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] bd7263b2-b996-4794-946b-2c28215574cb/bd7263b2-b996-4794-946b-2c28215574cb.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 870.956980] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f5b1d1fb-cc1c-4285-a519-c74370e78a2f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.985676] env[62385]: DEBUG oslo_vmware.api [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 870.985676] env[62385]: value = "task-1205958" [ 870.985676] env[62385]: _type = "Task" [ 870.985676] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.998374] env[62385]: DEBUG oslo_vmware.api [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205958, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.069183] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b5064658-1041-4627-8e74-39b389b439bf tempest-ServersAdminTestJSON-493781504 tempest-ServersAdminTestJSON-493781504-project-admin] Releasing lock "refresh_cache-b7a286a2-9a19-4ef1-8a46-6e2a04e22413" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.069687] env[62385]: DEBUG nova.compute.manager [None req-b5064658-1041-4627-8e74-39b389b439bf tempest-ServersAdminTestJSON-493781504 tempest-ServersAdminTestJSON-493781504-project-admin] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Inject network info {{(pid=62385) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7219}} [ 871.070124] env[62385]: DEBUG nova.compute.manager [None req-b5064658-1041-4627-8e74-39b389b439bf tempest-ServersAdminTestJSON-493781504 tempest-ServersAdminTestJSON-493781504-project-admin] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] network_info to inject: |[{"id": "2e49e027-be2a-4225-9a34-1e2f9cfb6022", "address": "fa:16:3e:cf:bd:fd", "network": {"id": "b0a3dea4-1925-4920-8a8e-9dc66733d1da", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1951021425-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ac0c8638f064daba042ded4de8c1cc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee20e439-fed9-490e-97dd-f3c886977ae1", "external-id": "nsx-vlan-transportzone-357", "segmentation_id": 357, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e49e027-be", "ovs_interfaceid": "2e49e027-be2a-4225-9a34-1e2f9cfb6022", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7220}} [ 871.075633] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-b5064658-1041-4627-8e74-39b389b439bf tempest-ServersAdminTestJSON-493781504 tempest-ServersAdminTestJSON-493781504-project-admin] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Reconfiguring VM instance to set the machine id {{(pid=62385) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1802}} [ 871.076649] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-73ba3194-3019-4632-8f75-52fa01274189 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.103060] env[62385]: DEBUG nova.network.neutron [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Successfully updated port: 0e719241-7a39-42dd-b37e-31cc75a4da74 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 871.105746] env[62385]: DEBUG oslo_vmware.api [None req-b5064658-1041-4627-8e74-39b389b439bf tempest-ServersAdminTestJSON-493781504 tempest-ServersAdminTestJSON-493781504-project-admin] Waiting for the task: (returnval){ [ 871.105746] env[62385]: value = "task-1205959" [ 871.105746] env[62385]: _type = "Task" [ 871.105746] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.125331] env[62385]: DEBUG oslo_vmware.api [None req-b5064658-1041-4627-8e74-39b389b439bf tempest-ServersAdminTestJSON-493781504 tempest-ServersAdminTestJSON-493781504-project-admin] Task: {'id': task-1205959, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.148862] env[62385]: DEBUG oslo_vmware.api [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': task-1205956, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.653607} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.153932] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] fbc71ada-d3b5-46f6-90a9-489c118d5126/fbc71ada-d3b5-46f6-90a9-489c118d5126.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 871.154351] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 871.155192] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a054158d-584f-4c2e-a2c7-83f5c31ef0e7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.166165] env[62385]: DEBUG oslo_vmware.api [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Waiting for the task: (returnval){ [ 871.166165] env[62385]: value = "task-1205960" [ 871.166165] env[62385]: _type = "Task" [ 871.166165] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.176871] env[62385]: DEBUG oslo_vmware.api [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': task-1205960, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.183431] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.561s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.183971] env[62385]: DEBUG nova.compute.manager [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 871.187868] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f2cd818c-6b68-4f26-b95e-028548504a6d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.165s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.188290] env[62385]: DEBUG nova.objects.instance [None req-f2cd818c-6b68-4f26-b95e-028548504a6d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lazy-loading 'resources' on Instance uuid 5a82fc98-e656-4476-b7dc-99466e9d6afd {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 871.282595] env[62385]: DEBUG nova.compute.manager [req-395e17ee-7cfa-4c45-bd2d-97c3d0a5a985 req-3cced874-48ec-4d7e-90a2-af7cdea5bc48 service nova] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Received event network-vif-plugged-0e719241-7a39-42dd-b37e-31cc75a4da74 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 871.282863] env[62385]: DEBUG oslo_concurrency.lockutils [req-395e17ee-7cfa-4c45-bd2d-97c3d0a5a985 req-3cced874-48ec-4d7e-90a2-af7cdea5bc48 service nova] Acquiring lock "d12ffea4-2275-4974-a5c6-b7fa365865b4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.283099] env[62385]: DEBUG oslo_concurrency.lockutils [req-395e17ee-7cfa-4c45-bd2d-97c3d0a5a985 req-3cced874-48ec-4d7e-90a2-af7cdea5bc48 service nova] Lock "d12ffea4-2275-4974-a5c6-b7fa365865b4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.283276] env[62385]: DEBUG oslo_concurrency.lockutils [req-395e17ee-7cfa-4c45-bd2d-97c3d0a5a985 req-3cced874-48ec-4d7e-90a2-af7cdea5bc48 service nova] Lock "d12ffea4-2275-4974-a5c6-b7fa365865b4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.283446] env[62385]: DEBUG nova.compute.manager [req-395e17ee-7cfa-4c45-bd2d-97c3d0a5a985 req-3cced874-48ec-4d7e-90a2-af7cdea5bc48 service nova] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] No waiting events found dispatching network-vif-plugged-0e719241-7a39-42dd-b37e-31cc75a4da74 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 871.283677] env[62385]: WARNING nova.compute.manager [req-395e17ee-7cfa-4c45-bd2d-97c3d0a5a985 req-3cced874-48ec-4d7e-90a2-af7cdea5bc48 service nova] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Received unexpected event network-vif-plugged-0e719241-7a39-42dd-b37e-31cc75a4da74 for instance with vm_state building and task_state spawning. [ 871.283793] env[62385]: DEBUG nova.compute.manager [req-395e17ee-7cfa-4c45-bd2d-97c3d0a5a985 req-3cced874-48ec-4d7e-90a2-af7cdea5bc48 service nova] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Received event network-changed-0e719241-7a39-42dd-b37e-31cc75a4da74 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 871.283946] env[62385]: DEBUG nova.compute.manager [req-395e17ee-7cfa-4c45-bd2d-97c3d0a5a985 req-3cced874-48ec-4d7e-90a2-af7cdea5bc48 service nova] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Refreshing instance network info cache due to event network-changed-0e719241-7a39-42dd-b37e-31cc75a4da74. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 871.284175] env[62385]: DEBUG oslo_concurrency.lockutils [req-395e17ee-7cfa-4c45-bd2d-97c3d0a5a985 req-3cced874-48ec-4d7e-90a2-af7cdea5bc48 service nova] Acquiring lock "refresh_cache-d12ffea4-2275-4974-a5c6-b7fa365865b4" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.284379] env[62385]: DEBUG oslo_concurrency.lockutils [req-395e17ee-7cfa-4c45-bd2d-97c3d0a5a985 req-3cced874-48ec-4d7e-90a2-af7cdea5bc48 service nova] Acquired lock "refresh_cache-d12ffea4-2275-4974-a5c6-b7fa365865b4" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.284553] env[62385]: DEBUG nova.network.neutron [req-395e17ee-7cfa-4c45-bd2d-97c3d0a5a985 req-3cced874-48ec-4d7e-90a2-af7cdea5bc48 service nova] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Refreshing network info cache for port 0e719241-7a39-42dd-b37e-31cc75a4da74 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 871.286571] env[62385]: DEBUG nova.network.neutron [None req-7766957d-6baa-406f-b861-20b7a4a54b1f tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Updating instance_info_cache with network_info: [{"id": "823867e4-d6b8-42d1-8597-3c27f6190ab1", "address": "fa:16:3e:56:b4:61", "network": {"id": "b563b70c-d3f1-4f82-97e9-defaf3f9d14b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-55206131-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "8eaa81326f084441a87ac05796958abb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac4015e0-e5e7-4b3f-8d8e-ef4501eea9aa", "external-id": "nsx-vlan-transportzone-132", "segmentation_id": 132, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap823867e4-d6", "ovs_interfaceid": "823867e4-d6b8-42d1-8597-3c27f6190ab1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.496651] env[62385]: DEBUG oslo_vmware.api [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205958, 'name': ReconfigVM_Task, 'duration_secs': 0.512525} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.496961] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Reconfigured VM instance instance-0000004f to attach disk [datastore2] bd7263b2-b996-4794-946b-2c28215574cb/bd7263b2-b996-4794-946b-2c28215574cb.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 871.497640] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-587b5e4f-0060-4231-82d7-743af71f2d26 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.504820] env[62385]: DEBUG oslo_vmware.api [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 871.504820] env[62385]: value = "task-1205961" [ 871.504820] env[62385]: _type = "Task" [ 871.504820] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.510020] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e650b13f-9669-4544-afb3-afe7fba05c51 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "fe0624a9-09a8-498b-bb3c-fda6cab92341" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.510020] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e650b13f-9669-4544-afb3-afe7fba05c51 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "fe0624a9-09a8-498b-bb3c-fda6cab92341" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.510020] env[62385]: DEBUG nova.compute.manager [None req-e650b13f-9669-4544-afb3-afe7fba05c51 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 871.510020] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fcefb0b-9b35-4bf4-afc5-2c82acc64746 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.517339] env[62385]: DEBUG nova.compute.manager [None req-e650b13f-9669-4544-afb3-afe7fba05c51 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62385) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 871.518035] env[62385]: DEBUG nova.objects.instance [None req-e650b13f-9669-4544-afb3-afe7fba05c51 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lazy-loading 'flavor' on Instance uuid fe0624a9-09a8-498b-bb3c-fda6cab92341 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 871.522654] env[62385]: DEBUG oslo_vmware.api [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205961, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.608091] env[62385]: DEBUG oslo_concurrency.lockutils [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "refresh_cache-d12ffea4-2275-4974-a5c6-b7fa365865b4" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.619927] env[62385]: DEBUG oslo_vmware.api [None req-b5064658-1041-4627-8e74-39b389b439bf tempest-ServersAdminTestJSON-493781504 tempest-ServersAdminTestJSON-493781504-project-admin] Task: {'id': task-1205959, 'name': ReconfigVM_Task, 'duration_secs': 0.201089} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.620080] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-b5064658-1041-4627-8e74-39b389b439bf tempest-ServersAdminTestJSON-493781504 tempest-ServersAdminTestJSON-493781504-project-admin] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Reconfigured VM instance to set the machine id {{(pid=62385) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1805}} [ 871.677275] env[62385]: DEBUG oslo_vmware.api [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': task-1205960, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.095399} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.677615] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 871.678484] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0d35868-286d-40e5-aceb-3fb8064b5fa8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.702916] env[62385]: DEBUG nova.compute.utils [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 871.713055] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] fbc71ada-d3b5-46f6-90a9-489c118d5126/fbc71ada-d3b5-46f6-90a9-489c118d5126.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 871.714023] env[62385]: DEBUG nova.compute.manager [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 871.714215] env[62385]: DEBUG nova.network.neutron [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 871.716913] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5d1b1dc8-0096-402d-9baf-122bb80da64b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.741457] env[62385]: DEBUG oslo_vmware.api [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Waiting for the task: (returnval){ [ 871.741457] env[62385]: value = "task-1205962" [ 871.741457] env[62385]: _type = "Task" [ 871.741457] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.751634] env[62385]: DEBUG oslo_vmware.api [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': task-1205962, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.777751] env[62385]: DEBUG nova.policy [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4ac552dc01fc4de2b97e36b41dae725a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '105150c170d74e268e82ab894e9bf0d5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 871.791112] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7766957d-6baa-406f-b861-20b7a4a54b1f tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Releasing lock "refresh_cache-5e5f43e8-ce45-45c7-a244-bfa724ea5e3c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.791726] env[62385]: DEBUG nova.objects.instance [None req-7766957d-6baa-406f-b861-20b7a4a54b1f tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Lazy-loading 'flavor' on Instance uuid 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 871.842394] env[62385]: DEBUG nova.network.neutron [req-395e17ee-7cfa-4c45-bd2d-97c3d0a5a985 req-3cced874-48ec-4d7e-90a2-af7cdea5bc48 service nova] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 871.986954] env[62385]: DEBUG nova.network.neutron [req-395e17ee-7cfa-4c45-bd2d-97c3d0a5a985 req-3cced874-48ec-4d7e-90a2-af7cdea5bc48 service nova] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.018842] env[62385]: DEBUG oslo_vmware.api [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205961, 'name': Rename_Task, 'duration_secs': 0.245599} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.019757] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 872.019757] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1f6650ac-ada1-40b2-ba6d-3f7215a4b2b0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.034025] env[62385]: DEBUG oslo_vmware.api [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 872.034025] env[62385]: value = "task-1205963" [ 872.034025] env[62385]: _type = "Task" [ 872.034025] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.034025] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e650b13f-9669-4544-afb3-afe7fba05c51 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 872.034025] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9ecb8fe4-80f4-43da-9d7f-879117b9a628 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.044647] env[62385]: DEBUG oslo_vmware.api [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205963, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.046543] env[62385]: DEBUG oslo_vmware.api [None req-e650b13f-9669-4544-afb3-afe7fba05c51 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 872.046543] env[62385]: value = "task-1205964" [ 872.046543] env[62385]: _type = "Task" [ 872.046543] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.066819] env[62385]: DEBUG oslo_vmware.api [None req-e650b13f-9669-4544-afb3-afe7fba05c51 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205964, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.068107] env[62385]: DEBUG nova.network.neutron [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Successfully created port: f5cfcc74-0f00-411e-8ee7-e26ec0a2fe7a {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 872.106203] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02889695-0e92-475f-9979-6bae895897e1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.115214] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc4ff30a-eb72-43ae-871c-bd56acad15cb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.149930] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07c03fcf-dbc0-4a3e-a611-e5fd9013579e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.159631] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d558843-256e-44ad-ad5b-322070a9ec5e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.176442] env[62385]: DEBUG nova.compute.provider_tree [None req-f2cd818c-6b68-4f26-b95e-028548504a6d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 872.215344] env[62385]: DEBUG nova.compute.manager [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 872.255031] env[62385]: DEBUG oslo_vmware.api [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': task-1205962, 'name': ReconfigVM_Task, 'duration_secs': 0.496081} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.255031] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Reconfigured VM instance instance-0000004d to attach disk [datastore1] fbc71ada-d3b5-46f6-90a9-489c118d5126/fbc71ada-d3b5-46f6-90a9-489c118d5126.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 872.255280] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4fbf20ec-6e68-4bf6-a618-9df34bc15414 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.263940] env[62385]: DEBUG oslo_vmware.api [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Waiting for the task: (returnval){ [ 872.263940] env[62385]: value = "task-1205965" [ 872.263940] env[62385]: _type = "Task" [ 872.263940] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.275901] env[62385]: DEBUG oslo_vmware.api [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': task-1205965, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.299376] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0344ec5-2c48-4e82-b070-2218db299276 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.326527] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7766957d-6baa-406f-b861-20b7a4a54b1f tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 872.326934] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6704a92e-ced4-40dc-ae4e-b49e26632e39 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.334850] env[62385]: DEBUG oslo_vmware.api [None req-7766957d-6baa-406f-b861-20b7a4a54b1f tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for the task: (returnval){ [ 872.334850] env[62385]: value = "task-1205966" [ 872.334850] env[62385]: _type = "Task" [ 872.334850] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.345576] env[62385]: DEBUG oslo_vmware.api [None req-7766957d-6baa-406f-b861-20b7a4a54b1f tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205966, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.489377] env[62385]: DEBUG oslo_concurrency.lockutils [req-395e17ee-7cfa-4c45-bd2d-97c3d0a5a985 req-3cced874-48ec-4d7e-90a2-af7cdea5bc48 service nova] Releasing lock "refresh_cache-d12ffea4-2275-4974-a5c6-b7fa365865b4" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.489869] env[62385]: DEBUG oslo_concurrency.lockutils [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquired lock "refresh_cache-d12ffea4-2275-4974-a5c6-b7fa365865b4" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.490050] env[62385]: DEBUG nova.network.neutron [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 872.544784] env[62385]: DEBUG oslo_vmware.api [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205963, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.561048] env[62385]: DEBUG oslo_vmware.api [None req-e650b13f-9669-4544-afb3-afe7fba05c51 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205964, 'name': PowerOffVM_Task, 'duration_secs': 0.360246} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.561048] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e650b13f-9669-4544-afb3-afe7fba05c51 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 872.561048] env[62385]: DEBUG nova.compute.manager [None req-e650b13f-9669-4544-afb3-afe7fba05c51 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 872.561775] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a984f189-5488-43a3-ac1a-0aeff512c53b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.584114] env[62385]: DEBUG oslo_vmware.rw_handles [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ada463-7601-a7d8-4deb-9352fca7de64/disk-0.vmdk. {{(pid=62385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 872.585092] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6179ef1e-0075-4d64-a06e-3486073197d9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.592954] env[62385]: DEBUG oslo_vmware.rw_handles [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ada463-7601-a7d8-4deb-9352fca7de64/disk-0.vmdk is in state: ready. {{(pid=62385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 872.593207] env[62385]: ERROR oslo_vmware.rw_handles [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ada463-7601-a7d8-4deb-9352fca7de64/disk-0.vmdk due to incomplete transfer. [ 872.593425] env[62385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-ea1da596-b70c-42ac-866c-0431aa180a5d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.602867] env[62385]: DEBUG oslo_vmware.rw_handles [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ada463-7601-a7d8-4deb-9352fca7de64/disk-0.vmdk. {{(pid=62385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 872.603203] env[62385]: DEBUG nova.virt.vmwareapi.images [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Uploaded image a204cf12-3e93-4857-af15-299e714bfa66 to the Glance image server {{(pid=62385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 872.605458] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Destroying the VM {{(pid=62385) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 872.606218] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-5979b006-7cf9-4230-af1b-286c668b57a4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.619036] env[62385]: DEBUG oslo_vmware.api [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Waiting for the task: (returnval){ [ 872.619036] env[62385]: value = "task-1205967" [ 872.619036] env[62385]: _type = "Task" [ 872.619036] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.634836] env[62385]: DEBUG oslo_vmware.api [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205967, 'name': Destroy_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.680379] env[62385]: DEBUG nova.scheduler.client.report [None req-f2cd818c-6b68-4f26-b95e-028548504a6d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 872.777347] env[62385]: DEBUG oslo_vmware.api [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': task-1205965, 'name': Rename_Task, 'duration_secs': 0.20387} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.777678] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 872.777978] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d1dbedf2-dea8-4e65-8b65-8741fcad78e2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.787051] env[62385]: DEBUG oslo_vmware.api [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Waiting for the task: (returnval){ [ 872.787051] env[62385]: value = "task-1205968" [ 872.787051] env[62385]: _type = "Task" [ 872.787051] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.796727] env[62385]: DEBUG oslo_vmware.api [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': task-1205968, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.850261] env[62385]: DEBUG oslo_vmware.api [None req-7766957d-6baa-406f-b861-20b7a4a54b1f tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205966, 'name': PowerOffVM_Task, 'duration_secs': 0.266095} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.851042] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7766957d-6baa-406f-b861-20b7a4a54b1f tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 872.861105] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-7766957d-6baa-406f-b861-20b7a4a54b1f tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Reconfiguring VM instance instance-0000004a to detach disk 2001 {{(pid=62385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 872.861636] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6a4c9a6c-3e23-4019-9703-642c542656a2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.892632] env[62385]: DEBUG oslo_vmware.api [None req-7766957d-6baa-406f-b861-20b7a4a54b1f tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for the task: (returnval){ [ 872.892632] env[62385]: value = "task-1205969" [ 872.892632] env[62385]: _type = "Task" [ 872.892632] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.906831] env[62385]: DEBUG oslo_vmware.api [None req-7766957d-6baa-406f-b861-20b7a4a54b1f tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205969, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.026859] env[62385]: DEBUG nova.network.neutron [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 873.049142] env[62385]: DEBUG oslo_vmware.api [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205963, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.076261] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e650b13f-9669-4544-afb3-afe7fba05c51 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "fe0624a9-09a8-498b-bb3c-fda6cab92341" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.567s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.132435] env[62385]: DEBUG oslo_vmware.api [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205967, 'name': Destroy_Task, 'duration_secs': 0.46366} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.133240] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Destroyed the VM [ 873.133624] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Deleting Snapshot of the VM instance {{(pid=62385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 873.134109] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-c0d367fd-2104-41b6-a780-b41f17f2517e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.144327] env[62385]: DEBUG oslo_vmware.api [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Waiting for the task: (returnval){ [ 873.144327] env[62385]: value = "task-1205970" [ 873.144327] env[62385]: _type = "Task" [ 873.144327] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.155937] env[62385]: DEBUG oslo_vmware.api [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205970, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.163681] env[62385]: INFO nova.compute.manager [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Rebuilding instance [ 873.185677] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f2cd818c-6b68-4f26-b95e-028548504a6d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.998s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.191389] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.162s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.195152] env[62385]: INFO nova.compute.claims [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 873.215903] env[62385]: DEBUG nova.compute.manager [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 873.216942] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-093b425d-c75f-460a-ba1d-56f0e18cf4c3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.229312] env[62385]: DEBUG nova.compute.manager [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 873.233255] env[62385]: INFO nova.scheduler.client.report [None req-f2cd818c-6b68-4f26-b95e-028548504a6d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Deleted allocations for instance 5a82fc98-e656-4476-b7dc-99466e9d6afd [ 873.270233] env[62385]: DEBUG nova.virt.hardware [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 873.270430] env[62385]: DEBUG nova.virt.hardware [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 873.270593] env[62385]: DEBUG nova.virt.hardware [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 873.270779] env[62385]: DEBUG nova.virt.hardware [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 873.270923] env[62385]: DEBUG nova.virt.hardware [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 873.271601] env[62385]: DEBUG nova.virt.hardware [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 873.272931] env[62385]: DEBUG nova.virt.hardware [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 873.272931] env[62385]: DEBUG nova.virt.hardware [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 873.272931] env[62385]: DEBUG nova.virt.hardware [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 873.273060] env[62385]: DEBUG nova.virt.hardware [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 873.273225] env[62385]: DEBUG nova.virt.hardware [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 873.274564] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-378347ca-82eb-4422-8b85-c497e75f369f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.278111] env[62385]: DEBUG nova.network.neutron [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Updating instance_info_cache with network_info: [{"id": "0e719241-7a39-42dd-b37e-31cc75a4da74", "address": "fa:16:3e:80:c9:48", "network": {"id": "afae6b07-d08c-413f-aa73-b545db1bb766", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1654965885-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "93111f5ba17c47de9691f90b99e03aa8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e719241-7a", "ovs_interfaceid": "0e719241-7a39-42dd-b37e-31cc75a4da74", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.286740] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccd0e895-3cc7-4944-8303-b783760fe369 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.309906] env[62385]: DEBUG oslo_vmware.api [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': task-1205968, 'name': PowerOnVM_Task} progress is 78%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.404995] env[62385]: DEBUG oslo_vmware.api [None req-7766957d-6baa-406f-b861-20b7a4a54b1f tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205969, 'name': ReconfigVM_Task, 'duration_secs': 0.368385} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.405326] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-7766957d-6baa-406f-b861-20b7a4a54b1f tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Reconfigured VM instance instance-0000004a to detach disk 2001 {{(pid=62385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 873.405525] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7766957d-6baa-406f-b861-20b7a4a54b1f tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 873.405801] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-387681b1-d217-4f5c-b7ba-bfbb7ab08445 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.413289] env[62385]: DEBUG oslo_vmware.api [None req-7766957d-6baa-406f-b861-20b7a4a54b1f tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for the task: (returnval){ [ 873.413289] env[62385]: value = "task-1205971" [ 873.413289] env[62385]: _type = "Task" [ 873.413289] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.421927] env[62385]: DEBUG oslo_vmware.api [None req-7766957d-6baa-406f-b861-20b7a4a54b1f tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205971, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.543864] env[62385]: DEBUG oslo_vmware.api [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205963, 'name': PowerOnVM_Task, 'duration_secs': 1.024874} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.545278] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 873.545278] env[62385]: INFO nova.compute.manager [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Took 11.59 seconds to spawn the instance on the hypervisor. [ 873.545278] env[62385]: DEBUG nova.compute.manager [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 873.545476] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cf1f453-2bf2-470e-9eeb-277d6397390a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.569101] env[62385]: DEBUG nova.compute.manager [req-7b2b802e-904d-4873-a6ef-0e4b15f465ff req-1d619961-b451-4374-acc4-652355e05450 service nova] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Received event network-vif-plugged-f5cfcc74-0f00-411e-8ee7-e26ec0a2fe7a {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 873.569424] env[62385]: DEBUG oslo_concurrency.lockutils [req-7b2b802e-904d-4873-a6ef-0e4b15f465ff req-1d619961-b451-4374-acc4-652355e05450 service nova] Acquiring lock "a58cb15f-e4df-4b1e-b09c-2f61dce7200f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.569608] env[62385]: DEBUG oslo_concurrency.lockutils [req-7b2b802e-904d-4873-a6ef-0e4b15f465ff req-1d619961-b451-4374-acc4-652355e05450 service nova] Lock "a58cb15f-e4df-4b1e-b09c-2f61dce7200f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.572044] env[62385]: DEBUG oslo_concurrency.lockutils [req-7b2b802e-904d-4873-a6ef-0e4b15f465ff req-1d619961-b451-4374-acc4-652355e05450 service nova] Lock "a58cb15f-e4df-4b1e-b09c-2f61dce7200f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.572044] env[62385]: DEBUG nova.compute.manager [req-7b2b802e-904d-4873-a6ef-0e4b15f465ff req-1d619961-b451-4374-acc4-652355e05450 service nova] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] No waiting events found dispatching network-vif-plugged-f5cfcc74-0f00-411e-8ee7-e26ec0a2fe7a {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 873.572044] env[62385]: WARNING nova.compute.manager [req-7b2b802e-904d-4873-a6ef-0e4b15f465ff req-1d619961-b451-4374-acc4-652355e05450 service nova] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Received unexpected event network-vif-plugged-f5cfcc74-0f00-411e-8ee7-e26ec0a2fe7a for instance with vm_state building and task_state spawning. [ 873.658476] env[62385]: DEBUG oslo_vmware.api [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205970, 'name': RemoveSnapshot_Task} progress is 16%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.676728] env[62385]: DEBUG nova.network.neutron [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Successfully updated port: f5cfcc74-0f00-411e-8ee7-e26ec0a2fe7a {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 873.734840] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 873.735163] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8e92c056-5fbf-4680-809d-9eda1f0444c8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.743373] env[62385]: DEBUG oslo_vmware.api [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 873.743373] env[62385]: value = "task-1205972" [ 873.743373] env[62385]: _type = "Task" [ 873.743373] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.743960] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f2cd818c-6b68-4f26-b95e-028548504a6d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "5a82fc98-e656-4476-b7dc-99466e9d6afd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.701s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.755160] env[62385]: DEBUG oslo_vmware.api [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205972, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.782195] env[62385]: DEBUG oslo_concurrency.lockutils [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Releasing lock "refresh_cache-d12ffea4-2275-4974-a5c6-b7fa365865b4" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.782480] env[62385]: DEBUG nova.compute.manager [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Instance network_info: |[{"id": "0e719241-7a39-42dd-b37e-31cc75a4da74", "address": "fa:16:3e:80:c9:48", "network": {"id": "afae6b07-d08c-413f-aa73-b545db1bb766", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1654965885-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "93111f5ba17c47de9691f90b99e03aa8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e719241-7a", "ovs_interfaceid": "0e719241-7a39-42dd-b37e-31cc75a4da74", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 873.783322] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:80:c9:48', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4adc8ed0-d11a-4510-9be0-b27c0da3a903', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0e719241-7a39-42dd-b37e-31cc75a4da74', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 873.792526] env[62385]: DEBUG oslo.service.loopingcall [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 873.793056] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 873.796987] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dd0eb8cf-3d00-4b6f-b082-6876f5f040e2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.819344] env[62385]: DEBUG oslo_vmware.api [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': task-1205968, 'name': PowerOnVM_Task, 'duration_secs': 0.882896} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.821113] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 873.821113] env[62385]: INFO nova.compute.manager [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Took 16.97 seconds to spawn the instance on the hypervisor. [ 873.821113] env[62385]: DEBUG nova.compute.manager [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 873.821472] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 873.821472] env[62385]: value = "task-1205973" [ 873.821472] env[62385]: _type = "Task" [ 873.821472] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.822308] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9fce787-8770-4e11-be0c-228cc247fcf3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.835144] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205973, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.923150] env[62385]: DEBUG oslo_vmware.api [None req-7766957d-6baa-406f-b861-20b7a4a54b1f tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205971, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.068698] env[62385]: INFO nova.compute.manager [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Took 28.47 seconds to build instance. [ 874.158243] env[62385]: DEBUG oslo_vmware.api [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205970, 'name': RemoveSnapshot_Task, 'duration_secs': 0.728169} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.158243] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Deleted Snapshot of the VM instance {{(pid=62385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 874.158243] env[62385]: INFO nova.compute.manager [None req-776178c9-cd8a-4fe2-a6df-c1df983b4673 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Took 15.37 seconds to snapshot the instance on the hypervisor. [ 874.179350] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "refresh_cache-a58cb15f-e4df-4b1e-b09c-2f61dce7200f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.179500] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquired lock "refresh_cache-a58cb15f-e4df-4b1e-b09c-2f61dce7200f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.179646] env[62385]: DEBUG nova.network.neutron [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 874.255312] env[62385]: DEBUG oslo_vmware.api [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205972, 'name': PowerOffVM_Task, 'duration_secs': 0.304576} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.258158] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 874.258437] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 874.259681] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44dc7ed8-0721-4838-83c8-94a426109726 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.263292] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a61cdaf1-3673-4b1a-b5bb-58fa8fda5a47 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "fe0624a9-09a8-498b-bb3c-fda6cab92341" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.263450] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a61cdaf1-3673-4b1a-b5bb-58fa8fda5a47 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "fe0624a9-09a8-498b-bb3c-fda6cab92341" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.263660] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a61cdaf1-3673-4b1a-b5bb-58fa8fda5a47 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "fe0624a9-09a8-498b-bb3c-fda6cab92341-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.263887] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a61cdaf1-3673-4b1a-b5bb-58fa8fda5a47 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "fe0624a9-09a8-498b-bb3c-fda6cab92341-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.264018] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a61cdaf1-3673-4b1a-b5bb-58fa8fda5a47 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "fe0624a9-09a8-498b-bb3c-fda6cab92341-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.267154] env[62385]: INFO nova.compute.manager [None req-a61cdaf1-3673-4b1a-b5bb-58fa8fda5a47 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Terminating instance [ 874.269558] env[62385]: DEBUG nova.compute.manager [None req-a61cdaf1-3673-4b1a-b5bb-58fa8fda5a47 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 874.269769] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a61cdaf1-3673-4b1a-b5bb-58fa8fda5a47 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 874.272842] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c87ba73-88f2-461b-9e59-0a5b9b1d6138 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.276178] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 874.279591] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d635a114-75f5-41c7-92f0-77f235914e56 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.291081] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a61cdaf1-3673-4b1a-b5bb-58fa8fda5a47 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 874.291400] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a1307fcf-8b55-42bc-baca-b0e70bbebb7a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.338505] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205973, 'name': CreateVM_Task, 'duration_secs': 0.503061} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.338685] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 874.339463] env[62385]: DEBUG oslo_concurrency.lockutils [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.339639] env[62385]: DEBUG oslo_concurrency.lockutils [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.339980] env[62385]: DEBUG oslo_concurrency.lockutils [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 874.340629] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-786ae8d2-60ac-4b9a-8ebf-baa1be42ca44 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.351609] env[62385]: INFO nova.compute.manager [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Took 43.09 seconds to build instance. [ 874.357641] env[62385]: DEBUG oslo_vmware.api [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 874.357641] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]526d4873-2b5f-677c-6201-d9223f45af7c" [ 874.357641] env[62385]: _type = "Task" [ 874.357641] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.368838] env[62385]: DEBUG oslo_vmware.api [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]526d4873-2b5f-677c-6201-d9223f45af7c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.389058] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a61cdaf1-3673-4b1a-b5bb-58fa8fda5a47 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 874.389324] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a61cdaf1-3673-4b1a-b5bb-58fa8fda5a47 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Deleting contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 874.389590] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a61cdaf1-3673-4b1a-b5bb-58fa8fda5a47 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Deleting the datastore file [datastore2] fe0624a9-09a8-498b-bb3c-fda6cab92341 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 874.390242] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 874.390242] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Deleting contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 874.390242] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Deleting the datastore file [datastore2] 33387505-c576-488b-8c9c-b064fe81a7d7 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 874.390487] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cb31ead9-d3f6-4a41-baf3-b042479de715 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.392534] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a0c913ad-62c6-4ca9-9144-e7f98dcf26dc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.401340] env[62385]: DEBUG oslo_vmware.api [None req-a61cdaf1-3673-4b1a-b5bb-58fa8fda5a47 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 874.401340] env[62385]: value = "task-1205977" [ 874.401340] env[62385]: _type = "Task" [ 874.401340] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.402784] env[62385]: DEBUG oslo_vmware.api [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 874.402784] env[62385]: value = "task-1205976" [ 874.402784] env[62385]: _type = "Task" [ 874.402784] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.424599] env[62385]: DEBUG oslo_vmware.api [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205976, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.424984] env[62385]: DEBUG oslo_vmware.api [None req-a61cdaf1-3673-4b1a-b5bb-58fa8fda5a47 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205977, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.436024] env[62385]: DEBUG oslo_vmware.api [None req-7766957d-6baa-406f-b861-20b7a4a54b1f tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205971, 'name': PowerOnVM_Task, 'duration_secs': 0.775877} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.436024] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7766957d-6baa-406f-b861-20b7a4a54b1f tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 874.436024] env[62385]: DEBUG nova.compute.manager [None req-7766957d-6baa-406f-b861-20b7a4a54b1f tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 874.436827] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50b018e5-96c1-433b-b4a6-a24778da2a75 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.570258] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a772bb4c-7e93-4a31-b916-e2f29814a9f9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.573787] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c20e39b2-ea00-4d40-8afd-636bf0163b0a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "bd7263b2-b996-4794-946b-2c28215574cb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.983s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.580141] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2422e8d-f37d-4d81-a09c-73cc0b8293b3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.613032] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6909b7c8-8a41-41b3-be72-c60eebc8969b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.621509] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c416a360-49b1-43a1-a00f-8201a6e91eff {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.637235] env[62385]: DEBUG nova.compute.provider_tree [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 874.713778] env[62385]: DEBUG nova.network.neutron [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 874.851116] env[62385]: DEBUG nova.network.neutron [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Updating instance_info_cache with network_info: [{"id": "f5cfcc74-0f00-411e-8ee7-e26ec0a2fe7a", "address": "fa:16:3e:18:18:ff", "network": {"id": "30c4e78b-6a64-47dc-aaaa-a210160c9f23", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1809587656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "105150c170d74e268e82ab894e9bf0d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10ff2092-e8eb-4768-ad4a-65a80560b447", "external-id": "nsx-vlan-transportzone-845", "segmentation_id": 845, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5cfcc74-0f", "ovs_interfaceid": "f5cfcc74-0f00-411e-8ee7-e26ec0a2fe7a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.853312] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3c4208ca-87ff-4b67-b06b-f44995993428 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Lock "fbc71ada-d3b5-46f6-90a9-489c118d5126" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.601s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.869657] env[62385]: DEBUG oslo_vmware.api [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]526d4873-2b5f-677c-6201-d9223f45af7c, 'name': SearchDatastore_Task, 'duration_secs': 0.015134} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.869998] env[62385]: DEBUG oslo_concurrency.lockutils [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.870268] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 874.870470] env[62385]: DEBUG oslo_concurrency.lockutils [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.870615] env[62385]: DEBUG oslo_concurrency.lockutils [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.870784] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 874.872689] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3446dcfc-0fdb-41bb-907b-fbc106bf3fb0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.882554] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 874.882756] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 874.883528] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a0c6b88c-1a15-4e6d-bdeb-a8089b947577 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.889808] env[62385]: DEBUG oslo_vmware.api [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 874.889808] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52035e72-2d55-a32c-fe91-fddf69576219" [ 874.889808] env[62385]: _type = "Task" [ 874.889808] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.899098] env[62385]: DEBUG oslo_vmware.api [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52035e72-2d55-a32c-fe91-fddf69576219, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.915976] env[62385]: DEBUG oslo_vmware.api [None req-a61cdaf1-3673-4b1a-b5bb-58fa8fda5a47 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1205977, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.169966} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.919014] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a61cdaf1-3673-4b1a-b5bb-58fa8fda5a47 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 874.919239] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a61cdaf1-3673-4b1a-b5bb-58fa8fda5a47 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Deleted contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 874.919434] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a61cdaf1-3673-4b1a-b5bb-58fa8fda5a47 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 874.919609] env[62385]: INFO nova.compute.manager [None req-a61cdaf1-3673-4b1a-b5bb-58fa8fda5a47 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Took 0.65 seconds to destroy the instance on the hypervisor. [ 874.919859] env[62385]: DEBUG oslo.service.loopingcall [None req-a61cdaf1-3673-4b1a-b5bb-58fa8fda5a47 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 874.920109] env[62385]: DEBUG oslo_vmware.api [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205976, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162181} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.920326] env[62385]: DEBUG nova.compute.manager [-] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 874.920424] env[62385]: DEBUG nova.network.neutron [-] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 874.922063] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 874.922259] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Deleted contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 874.922437] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 875.141160] env[62385]: DEBUG nova.scheduler.client.report [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 875.354039] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Releasing lock "refresh_cache-a58cb15f-e4df-4b1e-b09c-2f61dce7200f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.354039] env[62385]: DEBUG nova.compute.manager [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Instance network_info: |[{"id": "f5cfcc74-0f00-411e-8ee7-e26ec0a2fe7a", "address": "fa:16:3e:18:18:ff", "network": {"id": "30c4e78b-6a64-47dc-aaaa-a210160c9f23", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1809587656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "105150c170d74e268e82ab894e9bf0d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10ff2092-e8eb-4768-ad4a-65a80560b447", "external-id": "nsx-vlan-transportzone-845", "segmentation_id": 845, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5cfcc74-0f", "ovs_interfaceid": "f5cfcc74-0f00-411e-8ee7-e26ec0a2fe7a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 875.354415] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:18:18:ff', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '10ff2092-e8eb-4768-ad4a-65a80560b447', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f5cfcc74-0f00-411e-8ee7-e26ec0a2fe7a', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 875.362189] env[62385]: DEBUG oslo.service.loopingcall [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 875.362694] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 875.362961] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7cf61781-bb72-47bf-b6b6-0187d6aa85ca {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.385361] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 875.385361] env[62385]: value = "task-1205978" [ 875.385361] env[62385]: _type = "Task" [ 875.385361] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.395019] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205978, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.403840] env[62385]: DEBUG oslo_vmware.api [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52035e72-2d55-a32c-fe91-fddf69576219, 'name': SearchDatastore_Task, 'duration_secs': 0.025812} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.404830] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c3e734a3-660c-4e30-83ee-b9e1315f08e8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.412536] env[62385]: DEBUG oslo_vmware.api [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 875.412536] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ce02a0-ee4c-dd8c-a298-8177b0d867a2" [ 875.412536] env[62385]: _type = "Task" [ 875.412536] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.424927] env[62385]: DEBUG oslo_vmware.api [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ce02a0-ee4c-dd8c-a298-8177b0d867a2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.601294] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6040b7c9-146e-4782-b07a-776996b26e1b tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Acquiring lock "fbc71ada-d3b5-46f6-90a9-489c118d5126" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.601619] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6040b7c9-146e-4782-b07a-776996b26e1b tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Lock "fbc71ada-d3b5-46f6-90a9-489c118d5126" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.601977] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6040b7c9-146e-4782-b07a-776996b26e1b tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Acquiring lock "fbc71ada-d3b5-46f6-90a9-489c118d5126-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.602207] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6040b7c9-146e-4782-b07a-776996b26e1b tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Lock "fbc71ada-d3b5-46f6-90a9-489c118d5126-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.602406] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6040b7c9-146e-4782-b07a-776996b26e1b tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Lock "fbc71ada-d3b5-46f6-90a9-489c118d5126-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.606168] env[62385]: INFO nova.compute.manager [None req-6040b7c9-146e-4782-b07a-776996b26e1b tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Terminating instance [ 875.608848] env[62385]: DEBUG nova.compute.manager [None req-6040b7c9-146e-4782-b07a-776996b26e1b tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 875.609188] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-6040b7c9-146e-4782-b07a-776996b26e1b tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 875.610416] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-784d6e14-8168-424a-94f8-b3d0a2d6df4b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.622397] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6040b7c9-146e-4782-b07a-776996b26e1b tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 875.622857] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6761d7e7-b3f5-4b5f-a5d3-a8341a65be0f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.632729] env[62385]: DEBUG oslo_vmware.api [None req-6040b7c9-146e-4782-b07a-776996b26e1b tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Waiting for the task: (returnval){ [ 875.632729] env[62385]: value = "task-1205979" [ 875.632729] env[62385]: _type = "Task" [ 875.632729] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.645948] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.454s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.646569] env[62385]: DEBUG nova.compute.manager [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 875.650283] env[62385]: DEBUG oslo_vmware.api [None req-6040b7c9-146e-4782-b07a-776996b26e1b tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': task-1205979, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.650657] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d33d6bc6-7bf9-4779-9990-fdee06785b9d tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.400s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.650918] env[62385]: DEBUG nova.objects.instance [None req-d33d6bc6-7bf9-4779-9990-fdee06785b9d tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Lazy-loading 'resources' on Instance uuid 7f2fe830-e16d-4684-91a4-aab219468e77 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 875.674198] env[62385]: DEBUG nova.network.neutron [-] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.718460] env[62385]: DEBUG nova.compute.manager [req-ed702345-c69c-461e-a375-456fb78665b9 req-c4311827-c9a7-4b52-b325-3092a671226f service nova] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Received event network-changed-f5cfcc74-0f00-411e-8ee7-e26ec0a2fe7a {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 875.718672] env[62385]: DEBUG nova.compute.manager [req-ed702345-c69c-461e-a375-456fb78665b9 req-c4311827-c9a7-4b52-b325-3092a671226f service nova] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Refreshing instance network info cache due to event network-changed-f5cfcc74-0f00-411e-8ee7-e26ec0a2fe7a. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 875.718944] env[62385]: DEBUG oslo_concurrency.lockutils [req-ed702345-c69c-461e-a375-456fb78665b9 req-c4311827-c9a7-4b52-b325-3092a671226f service nova] Acquiring lock "refresh_cache-a58cb15f-e4df-4b1e-b09c-2f61dce7200f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.719378] env[62385]: DEBUG oslo_concurrency.lockutils [req-ed702345-c69c-461e-a375-456fb78665b9 req-c4311827-c9a7-4b52-b325-3092a671226f service nova] Acquired lock "refresh_cache-a58cb15f-e4df-4b1e-b09c-2f61dce7200f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.719818] env[62385]: DEBUG nova.network.neutron [req-ed702345-c69c-461e-a375-456fb78665b9 req-c4311827-c9a7-4b52-b325-3092a671226f service nova] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Refreshing network info cache for port f5cfcc74-0f00-411e-8ee7-e26ec0a2fe7a {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 875.744530] env[62385]: DEBUG nova.compute.manager [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 875.746866] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df07b478-f8d6-4411-ad64-0b257c38f79a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.895481] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205978, 'name': CreateVM_Task, 'duration_secs': 0.379152} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.895686] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 875.896405] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.896579] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.896920] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 875.897231] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69615fe1-2e94-4082-a40c-b89b5bc9bdba {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.902554] env[62385]: DEBUG oslo_vmware.api [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 875.902554] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]523f86f0-7aea-b06b-38f1-0331ebff8f05" [ 875.902554] env[62385]: _type = "Task" [ 875.902554] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.912644] env[62385]: DEBUG oslo_vmware.api [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]523f86f0-7aea-b06b-38f1-0331ebff8f05, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.921681] env[62385]: DEBUG oslo_vmware.api [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ce02a0-ee4c-dd8c-a298-8177b0d867a2, 'name': SearchDatastore_Task, 'duration_secs': 0.014586} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.921956] env[62385]: DEBUG oslo_concurrency.lockutils [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.922254] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] d12ffea4-2275-4974-a5c6-b7fa365865b4/d12ffea4-2275-4974-a5c6-b7fa365865b4.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 875.922518] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bbd57493-2418-46be-8896-70002285bd39 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.934499] env[62385]: DEBUG oslo_vmware.api [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 875.934499] env[62385]: value = "task-1205980" [ 875.934499] env[62385]: _type = "Task" [ 875.934499] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.936406] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6a258a04-6455-49dc-9069-3e3c3f39df66 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Acquiring lock "5e5f43e8-ce45-45c7-a244-bfa724ea5e3c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.936406] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6a258a04-6455-49dc-9069-3e3c3f39df66 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Lock "5e5f43e8-ce45-45c7-a244-bfa724ea5e3c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.936406] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6a258a04-6455-49dc-9069-3e3c3f39df66 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Acquiring lock "5e5f43e8-ce45-45c7-a244-bfa724ea5e3c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.936406] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6a258a04-6455-49dc-9069-3e3c3f39df66 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Lock "5e5f43e8-ce45-45c7-a244-bfa724ea5e3c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.936406] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6a258a04-6455-49dc-9069-3e3c3f39df66 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Lock "5e5f43e8-ce45-45c7-a244-bfa724ea5e3c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.941559] env[62385]: INFO nova.compute.manager [None req-6a258a04-6455-49dc-9069-3e3c3f39df66 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Terminating instance [ 875.944812] env[62385]: DEBUG nova.compute.manager [None req-6a258a04-6455-49dc-9069-3e3c3f39df66 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 875.944812] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-6a258a04-6455-49dc-9069-3e3c3f39df66 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 875.945730] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a32edf5-ebb4-45f0-8373-017771695404 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.951733] env[62385]: DEBUG oslo_vmware.api [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1205980, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.958789] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a258a04-6455-49dc-9069-3e3c3f39df66 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 875.959170] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bf553afa-690b-4201-8329-894a064a7709 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.962675] env[62385]: DEBUG nova.virt.hardware [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 875.962871] env[62385]: DEBUG nova.virt.hardware [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 875.963042] env[62385]: DEBUG nova.virt.hardware [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 875.963242] env[62385]: DEBUG nova.virt.hardware [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 875.963404] env[62385]: DEBUG nova.virt.hardware [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 875.963558] env[62385]: DEBUG nova.virt.hardware [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 875.963770] env[62385]: DEBUG nova.virt.hardware [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 875.963931] env[62385]: DEBUG nova.virt.hardware [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 875.964126] env[62385]: DEBUG nova.virt.hardware [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 875.964297] env[62385]: DEBUG nova.virt.hardware [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 875.964475] env[62385]: DEBUG nova.virt.hardware [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 875.965345] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-776729a8-85b3-4a9b-b536-a73b5c7bb5fa {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.974189] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-decfb3ef-20f6-47ed-88b4-064861ff01e3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.977993] env[62385]: DEBUG oslo_vmware.api [None req-6a258a04-6455-49dc-9069-3e3c3f39df66 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for the task: (returnval){ [ 875.977993] env[62385]: value = "task-1205981" [ 875.977993] env[62385]: _type = "Task" [ 875.977993] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.992105] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:74:68:57', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ee20e439-fed9-490e-97dd-f3c886977ae1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '72d846e7-a597-4eb7-87ea-0d4309ceb42b', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 876.000257] env[62385]: DEBUG oslo.service.loopingcall [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 876.001162] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 876.001313] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-13c5748a-05c0-4449-9070-ebbc53ff4757 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.019383] env[62385]: DEBUG oslo_vmware.api [None req-6a258a04-6455-49dc-9069-3e3c3f39df66 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205981, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.026199] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 876.026199] env[62385]: value = "task-1205982" [ 876.026199] env[62385]: _type = "Task" [ 876.026199] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.035205] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205982, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.145274] env[62385]: DEBUG oslo_vmware.api [None req-6040b7c9-146e-4782-b07a-776996b26e1b tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': task-1205979, 'name': PowerOffVM_Task, 'duration_secs': 0.239268} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.145786] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6040b7c9-146e-4782-b07a-776996b26e1b tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 876.145786] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-6040b7c9-146e-4782-b07a-776996b26e1b tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 876.146121] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-17b41e82-b06f-4442-ad8e-a2f0dedc6d1c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.151576] env[62385]: DEBUG nova.compute.utils [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 876.153209] env[62385]: DEBUG nova.compute.manager [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 876.154320] env[62385]: DEBUG nova.network.neutron [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 876.177625] env[62385]: INFO nova.compute.manager [-] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Took 1.26 seconds to deallocate network for instance. [ 876.230087] env[62385]: DEBUG nova.policy [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd5b491ab2da849b5a78a1dee48e1b185', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '79f1a45c9d2b439692c00a26ad59371d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 876.262840] env[62385]: INFO nova.compute.manager [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] instance snapshotting [ 876.268740] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbf23504-cc38-4ee5-869d-53b1a7c45eb4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.294775] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4620b9aa-e2ca-437d-a321-44143e9c1da9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.315687] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-6040b7c9-146e-4782-b07a-776996b26e1b tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 876.315687] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-6040b7c9-146e-4782-b07a-776996b26e1b tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 876.315798] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-6040b7c9-146e-4782-b07a-776996b26e1b tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Deleting the datastore file [datastore1] fbc71ada-d3b5-46f6-90a9-489c118d5126 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 876.319022] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cbaabfa8-2cb3-41b3-a47a-50560114e365 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.331197] env[62385]: DEBUG oslo_vmware.api [None req-6040b7c9-146e-4782-b07a-776996b26e1b tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Waiting for the task: (returnval){ [ 876.331197] env[62385]: value = "task-1205984" [ 876.331197] env[62385]: _type = "Task" [ 876.331197] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.340823] env[62385]: DEBUG oslo_vmware.api [None req-6040b7c9-146e-4782-b07a-776996b26e1b tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': task-1205984, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.413733] env[62385]: DEBUG oslo_vmware.api [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]523f86f0-7aea-b06b-38f1-0331ebff8f05, 'name': SearchDatastore_Task, 'duration_secs': 0.025998} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.416770] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.417078] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 876.417341] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.417490] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.418149] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 876.421555] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d35b0329-d2e3-401e-a712-46184d004aeb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.435180] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 876.435575] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 876.446187] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6cb9e784-70bb-46fa-a311-f7828a9b4f33 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.476493] env[62385]: DEBUG oslo_vmware.api [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 876.476493] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]522cd051-75a1-b95c-8908-695f0813d071" [ 876.476493] env[62385]: _type = "Task" [ 876.476493] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.476825] env[62385]: DEBUG oslo_vmware.api [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1205980, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.495858] env[62385]: DEBUG oslo_vmware.api [None req-6a258a04-6455-49dc-9069-3e3c3f39df66 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205981, 'name': PowerOffVM_Task, 'duration_secs': 0.219665} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.496131] env[62385]: DEBUG oslo_vmware.api [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]522cd051-75a1-b95c-8908-695f0813d071, 'name': SearchDatastore_Task, 'duration_secs': 0.021017} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.498781] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a258a04-6455-49dc-9069-3e3c3f39df66 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 876.499015] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-6a258a04-6455-49dc-9069-3e3c3f39df66 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 876.499849] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-148dd930-ab6e-48af-8ac8-8a60d620c30e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.501389] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11555f60-c6ff-4c9a-9da7-20e031b2f307 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.511517] env[62385]: DEBUG oslo_vmware.api [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 876.511517] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52be1d35-6123-e7d3-3bbf-c2b693188647" [ 876.511517] env[62385]: _type = "Task" [ 876.511517] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.522104] env[62385]: DEBUG oslo_vmware.api [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52be1d35-6123-e7d3-3bbf-c2b693188647, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.536034] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1205982, 'name': CreateVM_Task, 'duration_secs': 0.395347} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.536281] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 876.536979] env[62385]: DEBUG oslo_concurrency.lockutils [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.537156] env[62385]: DEBUG oslo_concurrency.lockutils [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.537537] env[62385]: DEBUG oslo_concurrency.lockutils [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 876.538676] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d23f32bb-8f4d-410b-80e7-7946357126f8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.544621] env[62385]: DEBUG nova.network.neutron [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Successfully created port: feaac30d-d553-4422-8c98-7a74ee001276 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 876.548445] env[62385]: DEBUG oslo_vmware.api [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 876.548445] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]520e86b8-d072-be51-7af1-f807ac6ab9a9" [ 876.548445] env[62385]: _type = "Task" [ 876.548445] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.559593] env[62385]: DEBUG oslo_vmware.api [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]520e86b8-d072-be51-7af1-f807ac6ab9a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.581161] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-6a258a04-6455-49dc-9069-3e3c3f39df66 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 876.581418] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-6a258a04-6455-49dc-9069-3e3c3f39df66 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 876.581609] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a258a04-6455-49dc-9069-3e3c3f39df66 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Deleting the datastore file [datastore1] 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 876.581885] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a71b3943-a02a-46be-b5ad-ee501730f1f3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.590346] env[62385]: DEBUG oslo_vmware.api [None req-6a258a04-6455-49dc-9069-3e3c3f39df66 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for the task: (returnval){ [ 876.590346] env[62385]: value = "task-1205986" [ 876.590346] env[62385]: _type = "Task" [ 876.590346] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.603550] env[62385]: DEBUG oslo_vmware.api [None req-6a258a04-6455-49dc-9069-3e3c3f39df66 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205986, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.604554] env[62385]: DEBUG nova.network.neutron [req-ed702345-c69c-461e-a375-456fb78665b9 req-c4311827-c9a7-4b52-b325-3092a671226f service nova] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Updated VIF entry in instance network info cache for port f5cfcc74-0f00-411e-8ee7-e26ec0a2fe7a. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 876.604795] env[62385]: DEBUG nova.network.neutron [req-ed702345-c69c-461e-a375-456fb78665b9 req-c4311827-c9a7-4b52-b325-3092a671226f service nova] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Updating instance_info_cache with network_info: [{"id": "f5cfcc74-0f00-411e-8ee7-e26ec0a2fe7a", "address": "fa:16:3e:18:18:ff", "network": {"id": "30c4e78b-6a64-47dc-aaaa-a210160c9f23", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1809587656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "105150c170d74e268e82ab894e9bf0d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10ff2092-e8eb-4768-ad4a-65a80560b447", "external-id": "nsx-vlan-transportzone-845", "segmentation_id": 845, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5cfcc74-0f", "ovs_interfaceid": "f5cfcc74-0f00-411e-8ee7-e26ec0a2fe7a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.629280] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf797984-3b6b-4ea6-a032-e2aaa073d775 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.637459] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd07024c-57ed-4456-97cf-a11fa1dfb1fa {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.669909] env[62385]: DEBUG nova.compute.manager [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 876.676535] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6936e984-3f6b-4e60-b706-f5d1f245474e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.683231] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cbd1028-c804-43c4-93e2-e25002be0f21 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.688127] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a61cdaf1-3673-4b1a-b5bb-58fa8fda5a47 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.698714] env[62385]: DEBUG nova.compute.provider_tree [None req-d33d6bc6-7bf9-4779-9990-fdee06785b9d tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 876.807292] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Creating Snapshot of the VM instance {{(pid=62385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 876.808345] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-b5a51ccf-3b4e-4d77-a270-3fd92abba153 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.817275] env[62385]: DEBUG oslo_vmware.api [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 876.817275] env[62385]: value = "task-1205987" [ 876.817275] env[62385]: _type = "Task" [ 876.817275] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.826966] env[62385]: DEBUG oslo_vmware.api [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205987, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.840619] env[62385]: DEBUG oslo_vmware.api [None req-6040b7c9-146e-4782-b07a-776996b26e1b tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': task-1205984, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.342776} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.841042] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-6040b7c9-146e-4782-b07a-776996b26e1b tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 876.841229] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-6040b7c9-146e-4782-b07a-776996b26e1b tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 876.841540] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-6040b7c9-146e-4782-b07a-776996b26e1b tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 876.841809] env[62385]: INFO nova.compute.manager [None req-6040b7c9-146e-4782-b07a-776996b26e1b tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Took 1.23 seconds to destroy the instance on the hypervisor. [ 876.842203] env[62385]: DEBUG oslo.service.loopingcall [None req-6040b7c9-146e-4782-b07a-776996b26e1b tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 876.842418] env[62385]: DEBUG nova.compute.manager [-] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 876.842543] env[62385]: DEBUG nova.network.neutron [-] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 876.956226] env[62385]: DEBUG oslo_vmware.api [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1205980, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.572967} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.956704] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] d12ffea4-2275-4974-a5c6-b7fa365865b4/d12ffea4-2275-4974-a5c6-b7fa365865b4.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 876.957091] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 876.957495] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-82fe6d9c-f521-4872-93a3-595b1af529ab {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.965097] env[62385]: DEBUG oslo_vmware.api [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 876.965097] env[62385]: value = "task-1205988" [ 876.965097] env[62385]: _type = "Task" [ 876.965097] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.973553] env[62385]: DEBUG oslo_vmware.api [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1205988, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.021676] env[62385]: DEBUG oslo_vmware.api [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52be1d35-6123-e7d3-3bbf-c2b693188647, 'name': SearchDatastore_Task, 'duration_secs': 0.011557} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.022080] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.022476] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] a58cb15f-e4df-4b1e-b09c-2f61dce7200f/a58cb15f-e4df-4b1e-b09c-2f61dce7200f.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 877.022860] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b8cfab8a-2aab-4a7e-b392-0cd5371ec411 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.030377] env[62385]: DEBUG oslo_vmware.api [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 877.030377] env[62385]: value = "task-1205989" [ 877.030377] env[62385]: _type = "Task" [ 877.030377] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.039328] env[62385]: DEBUG oslo_vmware.api [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205989, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.058702] env[62385]: DEBUG oslo_vmware.api [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]520e86b8-d072-be51-7af1-f807ac6ab9a9, 'name': SearchDatastore_Task, 'duration_secs': 0.013966} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.059236] env[62385]: DEBUG oslo_concurrency.lockutils [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.059605] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 877.062030] env[62385]: DEBUG oslo_concurrency.lockutils [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.062030] env[62385]: DEBUG oslo_concurrency.lockutils [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.062030] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 877.062030] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b7cb797d-a3bc-403d-aa1c-3bf128dc1c0a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.075450] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 877.075624] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 877.076442] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-98b8051b-46f9-4096-a83f-e01342b2331a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.082517] env[62385]: DEBUG oslo_vmware.api [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 877.082517] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]523b0037-cb24-6db0-a723-d495da3bc21f" [ 877.082517] env[62385]: _type = "Task" [ 877.082517] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.099087] env[62385]: DEBUG oslo_vmware.api [None req-6a258a04-6455-49dc-9069-3e3c3f39df66 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1205986, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.283336} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.103112] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a258a04-6455-49dc-9069-3e3c3f39df66 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 877.103112] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-6a258a04-6455-49dc-9069-3e3c3f39df66 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 877.103112] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-6a258a04-6455-49dc-9069-3e3c3f39df66 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 877.103112] env[62385]: INFO nova.compute.manager [None req-6a258a04-6455-49dc-9069-3e3c3f39df66 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Took 1.16 seconds to destroy the instance on the hypervisor. [ 877.103112] env[62385]: DEBUG oslo.service.loopingcall [None req-6a258a04-6455-49dc-9069-3e3c3f39df66 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 877.103112] env[62385]: DEBUG oslo_vmware.api [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]523b0037-cb24-6db0-a723-d495da3bc21f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.103558] env[62385]: DEBUG nova.compute.manager [-] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 877.103558] env[62385]: DEBUG nova.network.neutron [-] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 877.107836] env[62385]: DEBUG oslo_concurrency.lockutils [req-ed702345-c69c-461e-a375-456fb78665b9 req-c4311827-c9a7-4b52-b325-3092a671226f service nova] Releasing lock "refresh_cache-a58cb15f-e4df-4b1e-b09c-2f61dce7200f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.108128] env[62385]: DEBUG nova.compute.manager [req-ed702345-c69c-461e-a375-456fb78665b9 req-c4311827-c9a7-4b52-b325-3092a671226f service nova] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Received event network-vif-deleted-7a94d167-2bed-4857-b8ce-bdc920b1982a {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 877.202251] env[62385]: DEBUG nova.scheduler.client.report [None req-d33d6bc6-7bf9-4779-9990-fdee06785b9d tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 877.296077] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Acquiring lock "f2be792c-a723-4a69-b255-bbe7f0e8692d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.296374] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Lock "f2be792c-a723-4a69-b255-bbe7f0e8692d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.296600] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Acquiring lock "f2be792c-a723-4a69-b255-bbe7f0e8692d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.296789] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Lock "f2be792c-a723-4a69-b255-bbe7f0e8692d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.296966] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Lock "f2be792c-a723-4a69-b255-bbe7f0e8692d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.300074] env[62385]: INFO nova.compute.manager [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Terminating instance [ 877.301930] env[62385]: DEBUG nova.compute.manager [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 877.302144] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 877.303009] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f66983a3-3554-4d48-a9c8-1733b404ae1f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.311178] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 877.311435] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-58b939df-13bb-4da0-86f1-2bbdad98186c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.318311] env[62385]: DEBUG oslo_vmware.api [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Waiting for the task: (returnval){ [ 877.318311] env[62385]: value = "task-1205990" [ 877.318311] env[62385]: _type = "Task" [ 877.318311] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.332810] env[62385]: DEBUG oslo_vmware.api [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205990, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.333190] env[62385]: DEBUG oslo_vmware.api [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205987, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.475527] env[62385]: DEBUG oslo_vmware.api [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1205988, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072683} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.476235] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 877.477137] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed296aa7-6844-4bc3-a5e9-7da59cce90de {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.505679] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] d12ffea4-2275-4974-a5c6-b7fa365865b4/d12ffea4-2275-4974-a5c6-b7fa365865b4.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 877.505679] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6833362a-62f1-43dc-8899-ca3056ef8700 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.527141] env[62385]: DEBUG oslo_vmware.api [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 877.527141] env[62385]: value = "task-1205991" [ 877.527141] env[62385]: _type = "Task" [ 877.527141] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.540610] env[62385]: DEBUG nova.compute.manager [req-fd35868a-85be-4d87-a9b6-41daaafc8ebf req-e4c31fcf-818f-418a-b8aa-dfb7ccb17903 service nova] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Received event network-vif-deleted-2fdc640e-c457-46d4-8a9d-fe1f934bd29d {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 877.540807] env[62385]: INFO nova.compute.manager [req-fd35868a-85be-4d87-a9b6-41daaafc8ebf req-e4c31fcf-818f-418a-b8aa-dfb7ccb17903 service nova] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Neutron deleted interface 2fdc640e-c457-46d4-8a9d-fe1f934bd29d; detaching it from the instance and deleting it from the info cache [ 877.541115] env[62385]: DEBUG nova.network.neutron [req-fd35868a-85be-4d87-a9b6-41daaafc8ebf req-e4c31fcf-818f-418a-b8aa-dfb7ccb17903 service nova] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Updating instance_info_cache with network_info: [{"id": "2572a945-d7b9-47e5-8641-3718434e80ca", "address": "fa:16:3e:80:c2:26", "network": {"id": "0e19be42-16f3-43a9-9921-1f763804af82", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-846114175", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.182", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d68661a0197d4e3cacc06ea433ac706e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7edb7c08-2fae-4df5-9ec6-5ccf06d7e337", "external-id": "nsx-vlan-transportzone-309", "segmentation_id": 309, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2572a945-d7", "ovs_interfaceid": "2572a945-d7b9-47e5-8641-3718434e80ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "6820d526-f9c3-43cc-91db-538e9c3bd249", "address": "fa:16:3e:1a:d1:98", "network": {"id": "24d1af4a-afeb-449c-9ff6-367ea00a6763", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-990399095", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.204", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "d68661a0197d4e3cacc06ea433ac706e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c02dd284-ab80-451c-93eb-48c8360acb9c", "external-id": "nsx-vlan-transportzone-818", "segmentation_id": 818, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6820d526-f9", "ovs_interfaceid": "6820d526-f9c3-43cc-91db-538e9c3bd249", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.549902] env[62385]: DEBUG oslo_vmware.api [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205989, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.550208] env[62385]: DEBUG oslo_vmware.api [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1205991, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.595547] env[62385]: DEBUG oslo_vmware.api [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]523b0037-cb24-6db0-a723-d495da3bc21f, 'name': SearchDatastore_Task, 'duration_secs': 0.029966} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.596551] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31c830de-aa25-40f6-b327-280f4260d232 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.603884] env[62385]: DEBUG oslo_vmware.api [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 877.603884] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c57394-3ce3-7ff7-49c6-b61b0dc32bf8" [ 877.603884] env[62385]: _type = "Task" [ 877.603884] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.616940] env[62385]: DEBUG oslo_vmware.api [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c57394-3ce3-7ff7-49c6-b61b0dc32bf8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.679136] env[62385]: DEBUG nova.compute.manager [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 877.708851] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d33d6bc6-7bf9-4779-9990-fdee06785b9d tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.057s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.712314] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.083s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.712314] env[62385]: INFO nova.compute.claims [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 877.722050] env[62385]: DEBUG nova.virt.hardware [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 877.723096] env[62385]: DEBUG nova.virt.hardware [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 877.723096] env[62385]: DEBUG nova.virt.hardware [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 877.723096] env[62385]: DEBUG nova.virt.hardware [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 877.723465] env[62385]: DEBUG nova.virt.hardware [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 877.723619] env[62385]: DEBUG nova.virt.hardware [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 877.724017] env[62385]: DEBUG nova.virt.hardware [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 877.725187] env[62385]: DEBUG nova.virt.hardware [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 877.725187] env[62385]: DEBUG nova.virt.hardware [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 877.725307] env[62385]: DEBUG nova.virt.hardware [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 877.725845] env[62385]: DEBUG nova.virt.hardware [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 877.727067] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1139f155-dfb7-49ad-bc62-b26cd51f66b8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.742227] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92b9b1f6-9670-4e70-beed-4cb09fee3a85 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.747770] env[62385]: INFO nova.scheduler.client.report [None req-d33d6bc6-7bf9-4779-9990-fdee06785b9d tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Deleted allocations for instance 7f2fe830-e16d-4684-91a4-aab219468e77 [ 877.840077] env[62385]: DEBUG oslo_vmware.api [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205990, 'name': PowerOffVM_Task, 'duration_secs': 0.237777} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.844395] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 877.844395] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 877.848020] env[62385]: DEBUG oslo_vmware.api [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205987, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.848020] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fc55ed8e-7769-4d63-bbdf-a775a2ea2f46 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.921124] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 877.921297] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 877.921611] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Deleting the datastore file [datastore1] f2be792c-a723-4a69-b255-bbe7f0e8692d {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 877.921789] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2e410a9a-1843-4b64-90d0-58ade25b2413 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.931865] env[62385]: DEBUG oslo_vmware.api [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Waiting for the task: (returnval){ [ 877.931865] env[62385]: value = "task-1205993" [ 877.931865] env[62385]: _type = "Task" [ 877.931865] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.945780] env[62385]: DEBUG oslo_vmware.api [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205993, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.996459] env[62385]: DEBUG nova.compute.manager [req-f059dceb-4dd4-427d-9e1d-cef7f82ce284 req-577679bc-fba4-4ea5-be6e-927dd3402ca8 service nova] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Received event network-vif-plugged-feaac30d-d553-4422-8c98-7a74ee001276 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 877.996685] env[62385]: DEBUG oslo_concurrency.lockutils [req-f059dceb-4dd4-427d-9e1d-cef7f82ce284 req-577679bc-fba4-4ea5-be6e-927dd3402ca8 service nova] Acquiring lock "e9bf98d1-089a-4078-b2f8-eb77943731e5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.996898] env[62385]: DEBUG oslo_concurrency.lockutils [req-f059dceb-4dd4-427d-9e1d-cef7f82ce284 req-577679bc-fba4-4ea5-be6e-927dd3402ca8 service nova] Lock "e9bf98d1-089a-4078-b2f8-eb77943731e5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.997087] env[62385]: DEBUG oslo_concurrency.lockutils [req-f059dceb-4dd4-427d-9e1d-cef7f82ce284 req-577679bc-fba4-4ea5-be6e-927dd3402ca8 service nova] Lock "e9bf98d1-089a-4078-b2f8-eb77943731e5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.997267] env[62385]: DEBUG nova.compute.manager [req-f059dceb-4dd4-427d-9e1d-cef7f82ce284 req-577679bc-fba4-4ea5-be6e-927dd3402ca8 service nova] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] No waiting events found dispatching network-vif-plugged-feaac30d-d553-4422-8c98-7a74ee001276 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 877.997437] env[62385]: WARNING nova.compute.manager [req-f059dceb-4dd4-427d-9e1d-cef7f82ce284 req-577679bc-fba4-4ea5-be6e-927dd3402ca8 service nova] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Received unexpected event network-vif-plugged-feaac30d-d553-4422-8c98-7a74ee001276 for instance with vm_state building and task_state spawning. [ 878.042816] env[62385]: DEBUG oslo_vmware.api [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1205991, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.047406] env[62385]: DEBUG oslo_vmware.api [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205989, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.928075} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.047621] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] a58cb15f-e4df-4b1e-b09c-2f61dce7200f/a58cb15f-e4df-4b1e-b09c-2f61dce7200f.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 878.047862] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 878.048331] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-14071784-7837-4a02-b170-6a5ae936273c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.051552] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-04546be1-076d-4a80-81e5-e144f7309949 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.055724] env[62385]: DEBUG oslo_vmware.api [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 878.055724] env[62385]: value = "task-1205994" [ 878.055724] env[62385]: _type = "Task" [ 878.055724] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.062766] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb626d4f-081a-4a6c-b26b-a30cf48e409f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.079036] env[62385]: DEBUG oslo_vmware.api [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205994, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.103781] env[62385]: DEBUG nova.compute.manager [req-fd35868a-85be-4d87-a9b6-41daaafc8ebf req-e4c31fcf-818f-418a-b8aa-dfb7ccb17903 service nova] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Detach interface failed, port_id=2fdc640e-c457-46d4-8a9d-fe1f934bd29d, reason: Instance fbc71ada-d3b5-46f6-90a9-489c118d5126 could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 878.116699] env[62385]: DEBUG oslo_vmware.api [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c57394-3ce3-7ff7-49c6-b61b0dc32bf8, 'name': SearchDatastore_Task, 'duration_secs': 0.02103} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.116699] env[62385]: DEBUG oslo_concurrency.lockutils [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.117682] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 33387505-c576-488b-8c9c-b064fe81a7d7/33387505-c576-488b-8c9c-b064fe81a7d7.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 878.117682] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4fded8c7-c024-4281-916d-b2037f0e9896 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.131314] env[62385]: DEBUG oslo_vmware.api [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 878.131314] env[62385]: value = "task-1205995" [ 878.131314] env[62385]: _type = "Task" [ 878.131314] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.140805] env[62385]: DEBUG oslo_vmware.api [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205995, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.231675] env[62385]: DEBUG nova.network.neutron [-] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.257228] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d33d6bc6-7bf9-4779-9990-fdee06785b9d tempest-ServerShowV257Test-1246776851 tempest-ServerShowV257Test-1246776851-project-member] Lock "7f2fe830-e16d-4684-91a4-aab219468e77" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.380s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.328290] env[62385]: DEBUG oslo_vmware.api [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205987, 'name': CreateSnapshot_Task, 'duration_secs': 1.125662} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.328567] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Created Snapshot of the VM instance {{(pid=62385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 878.329422] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85e170b1-3724-4472-8de3-2cbcd2c9abbc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.443838] env[62385]: DEBUG oslo_vmware.api [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1205993, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.417368} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.443838] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 878.444067] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 878.444206] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 878.444376] env[62385]: INFO nova.compute.manager [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Took 1.14 seconds to destroy the instance on the hypervisor. [ 878.444629] env[62385]: DEBUG oslo.service.loopingcall [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 878.444829] env[62385]: DEBUG nova.compute.manager [-] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 878.444944] env[62385]: DEBUG nova.network.neutron [-] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 878.539837] env[62385]: DEBUG oslo_vmware.api [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1205991, 'name': ReconfigVM_Task, 'duration_secs': 0.688228} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.540619] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Reconfigured VM instance instance-00000050 to attach disk [datastore2] d12ffea4-2275-4974-a5c6-b7fa365865b4/d12ffea4-2275-4974-a5c6-b7fa365865b4.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 878.542000] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ff2cffef-6fab-43f9-b153-9e71c5c1dc78 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.550414] env[62385]: DEBUG oslo_vmware.api [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 878.550414] env[62385]: value = "task-1205996" [ 878.550414] env[62385]: _type = "Task" [ 878.550414] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.566274] env[62385]: DEBUG oslo_vmware.api [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1205996, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.567247] env[62385]: DEBUG nova.network.neutron [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Successfully updated port: feaac30d-d553-4422-8c98-7a74ee001276 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 878.573497] env[62385]: DEBUG oslo_vmware.api [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205994, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072806} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.573784] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 878.574640] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86984915-23e5-4f3e-9e6a-44c0884fbf2d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.602449] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] a58cb15f-e4df-4b1e-b09c-2f61dce7200f/a58cb15f-e4df-4b1e-b09c-2f61dce7200f.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 878.603464] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-64f397dc-fd5a-4d39-a0b0-16742301d055 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.626671] env[62385]: DEBUG oslo_vmware.api [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 878.626671] env[62385]: value = "task-1205997" [ 878.626671] env[62385]: _type = "Task" [ 878.626671] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.641179] env[62385]: DEBUG oslo_vmware.api [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205995, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.644510] env[62385]: DEBUG nova.network.neutron [-] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.645819] env[62385]: DEBUG oslo_vmware.api [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205997, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.715351] env[62385]: DEBUG nova.compute.manager [req-7662af87-40c1-4daf-bbca-fc116bb678e8 req-8eb0d17a-bf06-4584-bc1a-f20a8f46cb9f service nova] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Received event network-changed-feaac30d-d553-4422-8c98-7a74ee001276 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 878.715553] env[62385]: DEBUG nova.compute.manager [req-7662af87-40c1-4daf-bbca-fc116bb678e8 req-8eb0d17a-bf06-4584-bc1a-f20a8f46cb9f service nova] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Refreshing instance network info cache due to event network-changed-feaac30d-d553-4422-8c98-7a74ee001276. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 878.715773] env[62385]: DEBUG oslo_concurrency.lockutils [req-7662af87-40c1-4daf-bbca-fc116bb678e8 req-8eb0d17a-bf06-4584-bc1a-f20a8f46cb9f service nova] Acquiring lock "refresh_cache-e9bf98d1-089a-4078-b2f8-eb77943731e5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.716408] env[62385]: DEBUG oslo_concurrency.lockutils [req-7662af87-40c1-4daf-bbca-fc116bb678e8 req-8eb0d17a-bf06-4584-bc1a-f20a8f46cb9f service nova] Acquired lock "refresh_cache-e9bf98d1-089a-4078-b2f8-eb77943731e5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.716408] env[62385]: DEBUG nova.network.neutron [req-7662af87-40c1-4daf-bbca-fc116bb678e8 req-8eb0d17a-bf06-4584-bc1a-f20a8f46cb9f service nova] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Refreshing network info cache for port feaac30d-d553-4422-8c98-7a74ee001276 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 878.734734] env[62385]: INFO nova.compute.manager [-] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Took 1.63 seconds to deallocate network for instance. [ 878.849807] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Creating linked-clone VM from snapshot {{(pid=62385) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 878.854422] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-b4e2b53d-aa45-418d-8bef-10c943f711ce {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.866363] env[62385]: DEBUG oslo_vmware.api [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 878.866363] env[62385]: value = "task-1205998" [ 878.866363] env[62385]: _type = "Task" [ 878.866363] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.882314] env[62385]: DEBUG oslo_vmware.api [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205998, 'name': CloneVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.065483] env[62385]: DEBUG oslo_vmware.api [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1205996, 'name': Rename_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.072642] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Acquiring lock "refresh_cache-e9bf98d1-089a-4078-b2f8-eb77943731e5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.112763] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e6a991f-befd-45da-a25a-87f7a45f2965 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.121639] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-065f9125-f41c-4147-b93d-6d6ac9f181a4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.156524] env[62385]: INFO nova.compute.manager [-] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Took 2.31 seconds to deallocate network for instance. [ 879.163333] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f923eb87-5c60-4e46-8e6a-82deec470836 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.175704] env[62385]: DEBUG oslo_vmware.api [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205997, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.182493] env[62385]: DEBUG oslo_vmware.api [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205995, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.026782} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.184715] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 33387505-c576-488b-8c9c-b064fe81a7d7/33387505-c576-488b-8c9c-b064fe81a7d7.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 879.184949] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 879.185278] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-692b0f65-ca17-42a5-9bb7-77940ced9dd2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.188758] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aca10be-e943-40b4-8c7e-624e2aa0c4ee {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.208421] env[62385]: DEBUG nova.compute.provider_tree [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 879.210877] env[62385]: DEBUG oslo_vmware.api [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 879.210877] env[62385]: value = "task-1205999" [ 879.210877] env[62385]: _type = "Task" [ 879.210877] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.223532] env[62385]: DEBUG oslo_vmware.api [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205999, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.240954] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6a258a04-6455-49dc-9069-3e3c3f39df66 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.265084] env[62385]: DEBUG nova.network.neutron [req-7662af87-40c1-4daf-bbca-fc116bb678e8 req-8eb0d17a-bf06-4584-bc1a-f20a8f46cb9f service nova] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 879.359393] env[62385]: DEBUG nova.network.neutron [-] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.374070] env[62385]: DEBUG nova.network.neutron [req-7662af87-40c1-4daf-bbca-fc116bb678e8 req-8eb0d17a-bf06-4584-bc1a-f20a8f46cb9f service nova] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.380142] env[62385]: DEBUG oslo_vmware.api [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205998, 'name': CloneVM_Task} progress is 93%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.565149] env[62385]: DEBUG oslo_vmware.api [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1205996, 'name': Rename_Task, 'duration_secs': 0.715145} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.565623] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 879.565983] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-22942b9b-e7b0-41fa-9a98-0e95794676c9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.573666] env[62385]: DEBUG nova.compute.manager [req-e638568d-75c1-4f5a-810c-1261f82e40c1 req-63cbd9f7-7c6c-467c-b55b-48cda9cd0f40 service nova] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Received event network-vif-deleted-823867e4-d6b8-42d1-8597-3c27f6190ab1 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 879.573875] env[62385]: DEBUG nova.compute.manager [req-e638568d-75c1-4f5a-810c-1261f82e40c1 req-63cbd9f7-7c6c-467c-b55b-48cda9cd0f40 service nova] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Received event network-vif-deleted-2572a945-d7b9-47e5-8641-3718434e80ca {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 879.574062] env[62385]: DEBUG nova.compute.manager [req-e638568d-75c1-4f5a-810c-1261f82e40c1 req-63cbd9f7-7c6c-467c-b55b-48cda9cd0f40 service nova] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Received event network-vif-deleted-6820d526-f9c3-43cc-91db-538e9c3bd249 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 879.574242] env[62385]: DEBUG nova.compute.manager [req-e638568d-75c1-4f5a-810c-1261f82e40c1 req-63cbd9f7-7c6c-467c-b55b-48cda9cd0f40 service nova] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Received event network-vif-deleted-936e490f-ca20-418c-9b14-7e99db5ee2a9 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 879.576051] env[62385]: DEBUG oslo_vmware.api [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 879.576051] env[62385]: value = "task-1206000" [ 879.576051] env[62385]: _type = "Task" [ 879.576051] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.584814] env[62385]: DEBUG oslo_vmware.api [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206000, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.637737] env[62385]: DEBUG oslo_vmware.api [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1205997, 'name': ReconfigVM_Task, 'duration_secs': 0.608345} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.638017] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Reconfigured VM instance instance-00000051 to attach disk [datastore1] a58cb15f-e4df-4b1e-b09c-2f61dce7200f/a58cb15f-e4df-4b1e-b09c-2f61dce7200f.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 879.638644] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1970ee46-bf65-4cac-a55c-b6ed6b1f7c65 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.646271] env[62385]: DEBUG oslo_vmware.api [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 879.646271] env[62385]: value = "task-1206001" [ 879.646271] env[62385]: _type = "Task" [ 879.646271] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.655361] env[62385]: DEBUG oslo_vmware.api [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206001, 'name': Rename_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.677548] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6040b7c9-146e-4782-b07a-776996b26e1b tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.712733] env[62385]: DEBUG nova.scheduler.client.report [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 879.724611] env[62385]: DEBUG oslo_vmware.api [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1205999, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074806} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.724883] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 879.725682] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07b0b439-f461-4565-b4bc-a16f68c6db00 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.748039] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Reconfiguring VM instance instance-0000002c to attach disk [datastore2] 33387505-c576-488b-8c9c-b064fe81a7d7/33387505-c576-488b-8c9c-b064fe81a7d7.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 879.748571] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-937184d5-b822-4199-b0f0-01804ba05a60 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.770134] env[62385]: DEBUG oslo_vmware.api [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 879.770134] env[62385]: value = "task-1206002" [ 879.770134] env[62385]: _type = "Task" [ 879.770134] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.777964] env[62385]: DEBUG oslo_vmware.api [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206002, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.862200] env[62385]: INFO nova.compute.manager [-] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Took 1.42 seconds to deallocate network for instance. [ 879.880794] env[62385]: DEBUG oslo_concurrency.lockutils [req-7662af87-40c1-4daf-bbca-fc116bb678e8 req-8eb0d17a-bf06-4584-bc1a-f20a8f46cb9f service nova] Releasing lock "refresh_cache-e9bf98d1-089a-4078-b2f8-eb77943731e5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.881816] env[62385]: DEBUG oslo_vmware.api [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205998, 'name': CloneVM_Task} progress is 93%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.883154] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Acquired lock "refresh_cache-e9bf98d1-089a-4078-b2f8-eb77943731e5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.883345] env[62385]: DEBUG nova.network.neutron [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 880.086448] env[62385]: DEBUG oslo_vmware.api [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206000, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.156109] env[62385]: DEBUG oslo_vmware.api [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206001, 'name': Rename_Task, 'duration_secs': 0.259777} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.156397] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 880.156663] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-84a10111-8fc3-4ba2-8bdc-a7b03dda0ba0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.163208] env[62385]: DEBUG oslo_vmware.api [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 880.163208] env[62385]: value = "task-1206003" [ 880.163208] env[62385]: _type = "Task" [ 880.163208] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.171132] env[62385]: DEBUG oslo_vmware.api [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206003, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.220436] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.510s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.220984] env[62385]: DEBUG nova.compute.manager [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 880.223876] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2bb6b7b6-ebba-481c-ae8f-c30766e091ab tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.090s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.224068] env[62385]: DEBUG nova.objects.instance [None req-2bb6b7b6-ebba-481c-ae8f-c30766e091ab tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lazy-loading 'resources' on Instance uuid b73a31df-53c8-4550-bf75-5cf3b5aff86c {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 880.281350] env[62385]: DEBUG oslo_vmware.api [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206002, 'name': ReconfigVM_Task, 'duration_secs': 0.35553} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.281350] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Reconfigured VM instance instance-0000002c to attach disk [datastore2] 33387505-c576-488b-8c9c-b064fe81a7d7/33387505-c576-488b-8c9c-b064fe81a7d7.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 880.281539] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bc7f5c71-7355-4503-b704-14f023c50376 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.287978] env[62385]: DEBUG oslo_vmware.api [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 880.287978] env[62385]: value = "task-1206004" [ 880.287978] env[62385]: _type = "Task" [ 880.287978] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.296442] env[62385]: DEBUG oslo_vmware.api [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206004, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.374442] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.379424] env[62385]: DEBUG oslo_vmware.api [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205998, 'name': CloneVM_Task} progress is 94%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.414834] env[62385]: DEBUG nova.network.neutron [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 880.582297] env[62385]: DEBUG nova.network.neutron [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Updating instance_info_cache with network_info: [{"id": "feaac30d-d553-4422-8c98-7a74ee001276", "address": "fa:16:3e:b1:42:8a", "network": {"id": "72a59677-7309-4da6-969a-1ae3e009ac7e", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1714045793-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "79f1a45c9d2b439692c00a26ad59371d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfeaac30d-d5", "ovs_interfaceid": "feaac30d-d553-4422-8c98-7a74ee001276", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.589402] env[62385]: DEBUG oslo_vmware.api [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206000, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.676458] env[62385]: DEBUG oslo_vmware.api [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206003, 'name': PowerOnVM_Task, 'duration_secs': 0.51113} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.676835] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 880.677155] env[62385]: INFO nova.compute.manager [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Took 7.45 seconds to spawn the instance on the hypervisor. [ 880.677422] env[62385]: DEBUG nova.compute.manager [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 880.678519] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1518c8b-e73d-4bec-8538-38206fc218e9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.726715] env[62385]: DEBUG nova.compute.utils [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 880.728650] env[62385]: DEBUG nova.compute.manager [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Not allocating networking since 'none' was specified. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 880.798232] env[62385]: DEBUG oslo_vmware.api [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206004, 'name': Rename_Task, 'duration_secs': 0.156659} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.800920] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 880.801463] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d2ab022c-4798-4dc0-be08-90566f2d6be7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.808379] env[62385]: DEBUG oslo_vmware.api [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 880.808379] env[62385]: value = "task-1206005" [ 880.808379] env[62385]: _type = "Task" [ 880.808379] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.818649] env[62385]: DEBUG oslo_vmware.api [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206005, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.884305] env[62385]: DEBUG oslo_vmware.api [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205998, 'name': CloneVM_Task} progress is 94%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.051082] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08e6e691-a5b3-46f9-84d9-c5f89af87301 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.059232] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e4aacb4-9148-4b04-b9b0-958ea86070bd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.088892] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Releasing lock "refresh_cache-e9bf98d1-089a-4078-b2f8-eb77943731e5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.089072] env[62385]: DEBUG nova.compute.manager [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Instance network_info: |[{"id": "feaac30d-d553-4422-8c98-7a74ee001276", "address": "fa:16:3e:b1:42:8a", "network": {"id": "72a59677-7309-4da6-969a-1ae3e009ac7e", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1714045793-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "79f1a45c9d2b439692c00a26ad59371d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e445fb59-822c-4d7d-943b-c8e3bbaca62e", "external-id": "nsx-vlan-transportzone-258", "segmentation_id": 258, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfeaac30d-d5", "ovs_interfaceid": "feaac30d-d553-4422-8c98-7a74ee001276", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 881.092743] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b1:42:8a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e445fb59-822c-4d7d-943b-c8e3bbaca62e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'feaac30d-d553-4422-8c98-7a74ee001276', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 881.100249] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Creating folder: Project (79f1a45c9d2b439692c00a26ad59371d). Parent ref: group-v261107. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 881.100998] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19fe5b9c-9809-47a1-81d6-4f97d14f069d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.103787] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6f8ffa13-2172-427a-b7b5-4951409f85dc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.112133] env[62385]: DEBUG oslo_vmware.api [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206000, 'name': PowerOnVM_Task, 'duration_secs': 1.317926} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.113965] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 881.114208] env[62385]: INFO nova.compute.manager [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Took 11.48 seconds to spawn the instance on the hypervisor. [ 881.114397] env[62385]: DEBUG nova.compute.manager [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 881.115730] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd5be5fe-1f39-4942-bf4b-883eae38c145 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.119470] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72f4b8b3-884b-4733-be88-3c769a4b060c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.125163] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Created folder: Project (79f1a45c9d2b439692c00a26ad59371d) in parent group-v261107. [ 881.125163] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Creating folder: Instances. Parent ref: group-v261241. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 881.125163] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ed613ea5-366b-44a5-902c-3f700682b852 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.139466] env[62385]: DEBUG nova.compute.provider_tree [None req-2bb6b7b6-ebba-481c-ae8f-c30766e091ab tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 881.144021] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Created folder: Instances in parent group-v261241. [ 881.144021] env[62385]: DEBUG oslo.service.loopingcall [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 881.144021] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 881.144021] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-23a02f73-3b3e-4e79-80ab-e4e643c3d064 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.164134] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 881.164134] env[62385]: value = "task-1206008" [ 881.164134] env[62385]: _type = "Task" [ 881.164134] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.174832] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206008, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.200114] env[62385]: INFO nova.compute.manager [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Took 29.23 seconds to build instance. [ 881.232884] env[62385]: DEBUG nova.compute.manager [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 881.318731] env[62385]: DEBUG oslo_vmware.api [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206005, 'name': PowerOnVM_Task} progress is 98%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.380234] env[62385]: DEBUG oslo_vmware.api [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205998, 'name': CloneVM_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.646744] env[62385]: DEBUG nova.scheduler.client.report [None req-2bb6b7b6-ebba-481c-ae8f-c30766e091ab tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 881.652383] env[62385]: INFO nova.compute.manager [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Took 31.69 seconds to build instance. [ 881.675258] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206008, 'name': CreateVM_Task, 'duration_secs': 0.44721} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.675433] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 881.676113] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.676317] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.676638] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 881.677299] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e9e73d1-b26e-40f9-97fc-7e1858d6149e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.681743] env[62385]: DEBUG oslo_vmware.api [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Waiting for the task: (returnval){ [ 881.681743] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a8225a-ec5e-84af-a3c7-66aae8e91657" [ 881.681743] env[62385]: _type = "Task" [ 881.681743] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.689540] env[62385]: DEBUG oslo_vmware.api [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a8225a-ec5e-84af-a3c7-66aae8e91657, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.704416] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e11c8432-8e3e-494c-a817-ffac9144ca7b tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "a58cb15f-e4df-4b1e-b09c-2f61dce7200f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.747s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.818992] env[62385]: DEBUG oslo_vmware.api [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206005, 'name': PowerOnVM_Task, 'duration_secs': 0.590075} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.819302] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 881.819603] env[62385]: DEBUG nova.compute.manager [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 881.820280] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e576238-b3f4-40f5-9a12-6cb289b23146 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.880812] env[62385]: DEBUG oslo_vmware.api [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1205998, 'name': CloneVM_Task, 'duration_secs': 2.577676} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.882054] env[62385]: INFO nova.virt.vmwareapi.vmops [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Created linked-clone VM from snapshot [ 881.882786] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d691aec1-42d9-452e-a866-fee423212684 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.890605] env[62385]: DEBUG nova.virt.vmwareapi.images [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Uploading image aa5b9007-7bfe-4917-a311-121e4cac3174 {{(pid=62385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 881.911732] env[62385]: DEBUG oslo_vmware.rw_handles [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 881.911732] env[62385]: value = "vm-261240" [ 881.911732] env[62385]: _type = "VirtualMachine" [ 881.911732] env[62385]: }. {{(pid=62385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 881.912057] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-61d09c54-ea5f-4105-90c2-31de2f9adeff {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.919303] env[62385]: DEBUG oslo_vmware.rw_handles [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lease: (returnval){ [ 881.919303] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52da9677-4a3a-0af3-c3ae-60de2a41da2e" [ 881.919303] env[62385]: _type = "HttpNfcLease" [ 881.919303] env[62385]: } obtained for exporting VM: (result){ [ 881.919303] env[62385]: value = "vm-261240" [ 881.919303] env[62385]: _type = "VirtualMachine" [ 881.919303] env[62385]: }. {{(pid=62385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 881.919550] env[62385]: DEBUG oslo_vmware.api [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the lease: (returnval){ [ 881.919550] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52da9677-4a3a-0af3-c3ae-60de2a41da2e" [ 881.919550] env[62385]: _type = "HttpNfcLease" [ 881.919550] env[62385]: } to be ready. {{(pid=62385) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 881.926468] env[62385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 881.926468] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52da9677-4a3a-0af3-c3ae-60de2a41da2e" [ 881.926468] env[62385]: _type = "HttpNfcLease" [ 881.926468] env[62385]: } is initializing. {{(pid=62385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 882.153235] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2bb6b7b6-ebba-481c-ae8f-c30766e091ab tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.929s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.158913] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.368s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.162181] env[62385]: INFO nova.compute.claims [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 882.165371] env[62385]: DEBUG oslo_concurrency.lockutils [None req-19279955-944a-4e9a-be4c-a5da58cd5468 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "d12ffea4-2275-4974-a5c6-b7fa365865b4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.222s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.192231] env[62385]: INFO nova.scheduler.client.report [None req-2bb6b7b6-ebba-481c-ae8f-c30766e091ab tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Deleted allocations for instance b73a31df-53c8-4550-bf75-5cf3b5aff86c [ 882.199444] env[62385]: DEBUG oslo_vmware.api [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a8225a-ec5e-84af-a3c7-66aae8e91657, 'name': SearchDatastore_Task, 'duration_secs': 0.023461} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.203104] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.203262] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 882.203534] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.203724] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.203947] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 882.204484] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-519dd642-cd7e-4c66-b564-f979b9080177 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.216565] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 882.216765] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 882.219902] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b2cfa53-a9ea-4a45-a30d-20e6a07447c9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.224913] env[62385]: DEBUG oslo_vmware.api [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Waiting for the task: (returnval){ [ 882.224913] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5288cd98-5c66-2cbe-61c3-ed0700f9e93c" [ 882.224913] env[62385]: _type = "Task" [ 882.224913] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.233200] env[62385]: DEBUG oslo_vmware.api [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5288cd98-5c66-2cbe-61c3-ed0700f9e93c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.242543] env[62385]: DEBUG nova.compute.manager [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 882.270454] env[62385]: DEBUG nova.virt.hardware [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 882.270710] env[62385]: DEBUG nova.virt.hardware [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 882.270871] env[62385]: DEBUG nova.virt.hardware [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 882.271071] env[62385]: DEBUG nova.virt.hardware [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 882.271231] env[62385]: DEBUG nova.virt.hardware [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 882.271383] env[62385]: DEBUG nova.virt.hardware [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 882.271594] env[62385]: DEBUG nova.virt.hardware [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 882.271756] env[62385]: DEBUG nova.virt.hardware [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 882.271927] env[62385]: DEBUG nova.virt.hardware [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 882.272114] env[62385]: DEBUG nova.virt.hardware [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 882.272296] env[62385]: DEBUG nova.virt.hardware [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 882.273148] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fab0608f-c785-4dd6-a03d-605e0bcfce8d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.280393] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbf3d26a-5ca5-4935-a466-84082f1f636e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.294213] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Instance VIF info [] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 882.300056] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Creating folder: Project (de365429f1db49cba4646c7f8df80e3a). Parent ref: group-v261107. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 882.300344] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-164edd2e-501a-4c85-8574-e55c80781517 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.315381] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Created folder: Project (de365429f1db49cba4646c7f8df80e3a) in parent group-v261107. [ 882.315574] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Creating folder: Instances. Parent ref: group-v261244. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 882.316142] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-994f9689-a362-44c7-b96e-ca4c348cf4dd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.326288] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Created folder: Instances in parent group-v261244. [ 882.326558] env[62385]: DEBUG oslo.service.loopingcall [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 882.327214] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 882.327214] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-db232e22-4284-4b57-9278-3723f5139e1a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.346811] env[62385]: DEBUG oslo_concurrency.lockutils [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.350412] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 882.350412] env[62385]: value = "task-1206012" [ 882.350412] env[62385]: _type = "Task" [ 882.350412] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.358287] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206012, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.429968] env[62385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 882.429968] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52da9677-4a3a-0af3-c3ae-60de2a41da2e" [ 882.429968] env[62385]: _type = "HttpNfcLease" [ 882.429968] env[62385]: } is ready. {{(pid=62385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 882.430450] env[62385]: DEBUG oslo_vmware.rw_handles [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 882.430450] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52da9677-4a3a-0af3-c3ae-60de2a41da2e" [ 882.430450] env[62385]: _type = "HttpNfcLease" [ 882.430450] env[62385]: }. {{(pid=62385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 882.432985] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1730fd3b-f6b4-4ba5-b939-d4dcb04435cd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.441212] env[62385]: DEBUG oslo_vmware.rw_handles [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528864e4-6de8-fe59-acb2-32151aebccca/disk-0.vmdk from lease info. {{(pid=62385) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 882.441212] env[62385]: DEBUG oslo_vmware.rw_handles [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528864e4-6de8-fe59-acb2-32151aebccca/disk-0.vmdk for reading. {{(pid=62385) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 882.497667] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ea700dc1-32d8-40eb-8bd2-acaa574e2aea tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "d12ffea4-2275-4974-a5c6-b7fa365865b4" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.497956] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ea700dc1-32d8-40eb-8bd2-acaa574e2aea tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "d12ffea4-2275-4974-a5c6-b7fa365865b4" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.498206] env[62385]: DEBUG nova.compute.manager [None req-ea700dc1-32d8-40eb-8bd2-acaa574e2aea tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 882.499607] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e897dcc9-35d2-47e9-a892-5d391b9b813f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.508469] env[62385]: DEBUG nova.compute.manager [None req-ea700dc1-32d8-40eb-8bd2-acaa574e2aea tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62385) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 882.509098] env[62385]: DEBUG nova.objects.instance [None req-ea700dc1-32d8-40eb-8bd2-acaa574e2aea tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lazy-loading 'flavor' on Instance uuid d12ffea4-2275-4974-a5c6-b7fa365865b4 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 882.545299] env[62385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-a5bced47-22df-4c45-9923-19c5ad37c35e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.708114] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2bb6b7b6-ebba-481c-ae8f-c30766e091ab tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "b73a31df-53c8-4550-bf75-5cf3b5aff86c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.657s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.735773] env[62385]: DEBUG oslo_vmware.api [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5288cd98-5c66-2cbe-61c3-ed0700f9e93c, 'name': SearchDatastore_Task, 'duration_secs': 0.016196} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.736594] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-709032ab-ffe6-4d32-8d43-a62133d789eb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.742281] env[62385]: DEBUG oslo_vmware.api [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Waiting for the task: (returnval){ [ 882.742281] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521f06be-2b92-da0b-443c-65c180643f0e" [ 882.742281] env[62385]: _type = "Task" [ 882.742281] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.749648] env[62385]: DEBUG oslo_vmware.api [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521f06be-2b92-da0b-443c-65c180643f0e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.859060] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206012, 'name': CreateVM_Task, 'duration_secs': 0.281987} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.859255] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 882.859694] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.859859] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.860262] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 882.860567] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c67152bb-df35-4f06-adba-a569c882cf79 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.865145] env[62385]: DEBUG oslo_vmware.api [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Waiting for the task: (returnval){ [ 882.865145] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a51309-af8e-dcac-e1c9-aa0611aae06d" [ 882.865145] env[62385]: _type = "Task" [ 882.865145] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.873519] env[62385]: DEBUG oslo_vmware.api [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a51309-af8e-dcac-e1c9-aa0611aae06d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.015422] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea700dc1-32d8-40eb-8bd2-acaa574e2aea tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 883.015885] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-99f2f034-fdbf-4a51-a178-20d48ae15047 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.023051] env[62385]: DEBUG oslo_vmware.api [None req-ea700dc1-32d8-40eb-8bd2-acaa574e2aea tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 883.023051] env[62385]: value = "task-1206013" [ 883.023051] env[62385]: _type = "Task" [ 883.023051] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.034042] env[62385]: DEBUG oslo_vmware.api [None req-ea700dc1-32d8-40eb-8bd2-acaa574e2aea tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206013, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.143737] env[62385]: DEBUG nova.compute.manager [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Stashing vm_state: active {{(pid=62385) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 883.257129] env[62385]: DEBUG oslo_vmware.api [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521f06be-2b92-da0b-443c-65c180643f0e, 'name': SearchDatastore_Task, 'duration_secs': 0.022152} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.257637] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.257981] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] e9bf98d1-089a-4078-b2f8-eb77943731e5/e9bf98d1-089a-4078-b2f8-eb77943731e5.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 883.258377] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cbe30fad-8eec-4693-9f04-6e4cfd6f42df {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.269528] env[62385]: DEBUG oslo_vmware.api [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Waiting for the task: (returnval){ [ 883.269528] env[62385]: value = "task-1206014" [ 883.269528] env[62385]: _type = "Task" [ 883.269528] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.282092] env[62385]: DEBUG oslo_vmware.api [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Task: {'id': task-1206014, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.321636] env[62385]: INFO nova.compute.manager [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Rebuilding instance [ 883.384256] env[62385]: DEBUG oslo_vmware.api [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a51309-af8e-dcac-e1c9-aa0611aae06d, 'name': SearchDatastore_Task, 'duration_secs': 0.012781} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.384872] env[62385]: DEBUG nova.compute.manager [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 883.385291] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.385419] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 883.385613] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.385753] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.385933] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 883.386725] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf7ce04e-41b3-438f-bfd1-345f41f6f196 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.394869] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cb355998-7fad-4f5d-ae2e-b00f0b509600 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.402812] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 883.403096] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 883.404198] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae055f5a-e7ce-4690-b60f-2fe588e3904e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.409979] env[62385]: DEBUG oslo_vmware.api [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Waiting for the task: (returnval){ [ 883.409979] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52b98413-c0c4-c427-49b6-ee3ac331d97d" [ 883.409979] env[62385]: _type = "Task" [ 883.409979] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.425319] env[62385]: DEBUG oslo_vmware.api [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52b98413-c0c4-c427-49b6-ee3ac331d97d, 'name': SearchDatastore_Task, 'duration_secs': 0.010638} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.429062] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dd7b083f-b0d8-4cb0-9259-bb8ce69c5f1f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.435367] env[62385]: DEBUG oslo_vmware.api [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Waiting for the task: (returnval){ [ 883.435367] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52b79b54-6d3c-0858-2d5b-cfba1f595ffa" [ 883.435367] env[62385]: _type = "Task" [ 883.435367] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.447731] env[62385]: DEBUG oslo_vmware.api [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52b79b54-6d3c-0858-2d5b-cfba1f595ffa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.537420] env[62385]: DEBUG oslo_vmware.api [None req-ea700dc1-32d8-40eb-8bd2-acaa574e2aea tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206013, 'name': PowerOffVM_Task, 'duration_secs': 0.261082} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.537763] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea700dc1-32d8-40eb-8bd2-acaa574e2aea tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 883.537970] env[62385]: DEBUG nova.compute.manager [None req-ea700dc1-32d8-40eb-8bd2-acaa574e2aea tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 883.539545] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5de6182b-abf5-4c95-92a1-9955dc5631eb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.576275] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08584cb2-ecdb-40c7-bb4f-e197c57642a6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.586586] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86ab7356-3be0-40ea-83e4-f26df5219064 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.626118] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dfb836e-5095-4090-9a57-614df71915f6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.632700] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "9b2f8292-9d89-407e-96c5-195ee398cc7b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.632998] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "9b2f8292-9d89-407e-96c5-195ee398cc7b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.642240] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa947087-9598-4ff0-9507-64238ce6fed3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.667969] env[62385]: DEBUG nova.compute.provider_tree [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 883.670614] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.782955] env[62385]: DEBUG oslo_vmware.api [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Task: {'id': task-1206014, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.906327] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 883.906642] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bea49f7f-0810-4e4b-8667-b4a2554d612e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.918833] env[62385]: DEBUG oslo_vmware.api [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 883.918833] env[62385]: value = "task-1206015" [ 883.918833] env[62385]: _type = "Task" [ 883.918833] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.928645] env[62385]: DEBUG oslo_vmware.api [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206015, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.946642] env[62385]: DEBUG oslo_vmware.api [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52b79b54-6d3c-0858-2d5b-cfba1f595ffa, 'name': SearchDatastore_Task, 'duration_secs': 0.012562} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.946977] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.947243] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66/4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 883.947528] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f273b4e1-8716-4547-beaf-868ca6ffacb5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.955986] env[62385]: DEBUG oslo_vmware.api [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Waiting for the task: (returnval){ [ 883.955986] env[62385]: value = "task-1206016" [ 883.955986] env[62385]: _type = "Task" [ 883.955986] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.965113] env[62385]: DEBUG oslo_vmware.api [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Task: {'id': task-1206016, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.060143] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ea700dc1-32d8-40eb-8bd2-acaa574e2aea tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "d12ffea4-2275-4974-a5c6-b7fa365865b4" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.562s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.137064] env[62385]: DEBUG nova.compute.manager [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 884.173489] env[62385]: DEBUG nova.scheduler.client.report [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 884.283197] env[62385]: DEBUG oslo_vmware.api [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Task: {'id': task-1206014, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.571785} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.283557] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] e9bf98d1-089a-4078-b2f8-eb77943731e5/e9bf98d1-089a-4078-b2f8-eb77943731e5.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 884.283878] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 884.284229] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cadc7d9f-7f03-4ca3-bedf-22dfcf17424b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.294809] env[62385]: DEBUG oslo_vmware.api [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Waiting for the task: (returnval){ [ 884.294809] env[62385]: value = "task-1206017" [ 884.294809] env[62385]: _type = "Task" [ 884.294809] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.307624] env[62385]: DEBUG oslo_vmware.api [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Task: {'id': task-1206017, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.429999] env[62385]: DEBUG oslo_vmware.api [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206015, 'name': PowerOffVM_Task, 'duration_secs': 0.254165} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.430468] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 884.430692] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 884.431544] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-689e0e00-dc56-46ce-aae7-ada14ff7f0c5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.439901] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 884.440276] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-35ba98da-4364-4796-ad2e-780dbff0d857 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.466552] env[62385]: DEBUG oslo_vmware.api [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Task: {'id': task-1206016, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.518287] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 884.518397] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Deleting contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 884.518587] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Deleting the datastore file [datastore2] 33387505-c576-488b-8c9c-b064fe81a7d7 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 884.518855] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-64b10420-22b6-42a0-8f97-de9437172ec0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.530420] env[62385]: DEBUG oslo_vmware.api [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 884.530420] env[62385]: value = "task-1206019" [ 884.530420] env[62385]: _type = "Task" [ 884.530420] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.541182] env[62385]: DEBUG oslo_vmware.api [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206019, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.663296] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.678881] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.521s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.679504] env[62385]: DEBUG nova.compute.manager [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 884.682391] env[62385]: DEBUG oslo_concurrency.lockutils [None req-032bafda-a74f-4b9e-8041-96a5dea4e54f tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.539s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.686049] env[62385]: DEBUG nova.objects.instance [None req-032bafda-a74f-4b9e-8041-96a5dea4e54f tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Lazy-loading 'resources' on Instance uuid 699a57a9-5a1a-4cd1-8449-723400d9caf4 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 884.806487] env[62385]: DEBUG oslo_vmware.api [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Task: {'id': task-1206017, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.093805} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.806872] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 884.807670] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60d85865-a812-4811-b1f2-a5f769538706 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.831413] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] e9bf98d1-089a-4078-b2f8-eb77943731e5/e9bf98d1-089a-4078-b2f8-eb77943731e5.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 884.831702] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0d07027c-dcfc-4dec-a205-42c5efc4259a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.853251] env[62385]: DEBUG oslo_vmware.api [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Waiting for the task: (returnval){ [ 884.853251] env[62385]: value = "task-1206020" [ 884.853251] env[62385]: _type = "Task" [ 884.853251] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.865278] env[62385]: DEBUG oslo_vmware.api [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Task: {'id': task-1206020, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.968681] env[62385]: DEBUG oslo_vmware.api [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Task: {'id': task-1206016, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.772888} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.969016] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66/4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 884.969922] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 884.969922] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2bc86307-d3c2-4340-ba4c-8b7acc3a2f2d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.977546] env[62385]: DEBUG oslo_vmware.api [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Waiting for the task: (returnval){ [ 884.977546] env[62385]: value = "task-1206021" [ 884.977546] env[62385]: _type = "Task" [ 884.977546] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.987509] env[62385]: DEBUG oslo_vmware.api [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Task: {'id': task-1206021, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.041824] env[62385]: DEBUG oslo_vmware.api [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206019, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.43967} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.042177] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 885.042402] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Deleted contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 885.042592] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 885.188543] env[62385]: DEBUG nova.compute.utils [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 885.190383] env[62385]: DEBUG nova.compute.manager [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Not allocating networking since 'none' was specified. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 885.363462] env[62385]: DEBUG oslo_vmware.api [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Task: {'id': task-1206020, 'name': ReconfigVM_Task, 'duration_secs': 0.406907} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.366157] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Reconfigured VM instance instance-00000052 to attach disk [datastore1] e9bf98d1-089a-4078-b2f8-eb77943731e5/e9bf98d1-089a-4078-b2f8-eb77943731e5.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 885.367740] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-874f6a32-a338-4c4d-9ea2-174ad15035d4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.375356] env[62385]: DEBUG oslo_vmware.api [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Waiting for the task: (returnval){ [ 885.375356] env[62385]: value = "task-1206022" [ 885.375356] env[62385]: _type = "Task" [ 885.375356] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.388896] env[62385]: DEBUG oslo_vmware.api [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Task: {'id': task-1206022, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.487839] env[62385]: DEBUG oslo_vmware.api [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Task: {'id': task-1206021, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.111015} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.488151] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 885.489014] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7e6f7fe-2c83-464e-ace2-226032cfe38e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.505539] env[62385]: INFO nova.compute.manager [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Rebuilding instance [ 885.518779] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66/4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 885.520421] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-daa1ec64-68ee-4f0e-8492-8f75d88f1caf {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.536056] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b22555ec-be85-4db5-a340-4508c193beb4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.555195] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db2a8a22-378e-4a56-aef0-ccfca8da9a80 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.558880] env[62385]: DEBUG oslo_vmware.api [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Waiting for the task: (returnval){ [ 885.558880] env[62385]: value = "task-1206023" [ 885.558880] env[62385]: _type = "Task" [ 885.558880] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.591019] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-883d11b3-3ae1-46eb-b753-efd62f32785b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.599475] env[62385]: DEBUG oslo_vmware.api [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Task: {'id': task-1206023, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.600130] env[62385]: DEBUG nova.compute.manager [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 885.601257] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14275526-2f9c-4069-a471-f336a40c6792 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.607292] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24a469c0-7022-4b05-8ed7-095e2851a305 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.627639] env[62385]: DEBUG nova.compute.provider_tree [None req-032bafda-a74f-4b9e-8041-96a5dea4e54f tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 885.693096] env[62385]: DEBUG nova.compute.manager [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 885.886513] env[62385]: DEBUG oslo_vmware.api [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Task: {'id': task-1206022, 'name': Rename_Task, 'duration_secs': 0.191577} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.887311] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 885.887742] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5ef0d090-cba7-4a97-a334-0b5719f92ab4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.895566] env[62385]: DEBUG oslo_vmware.api [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Waiting for the task: (returnval){ [ 885.895566] env[62385]: value = "task-1206024" [ 885.895566] env[62385]: _type = "Task" [ 885.895566] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.904737] env[62385]: DEBUG oslo_vmware.api [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Task: {'id': task-1206024, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.073251] env[62385]: DEBUG oslo_vmware.api [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Task: {'id': task-1206023, 'name': ReconfigVM_Task, 'duration_secs': 0.336117} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.073543] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Reconfigured VM instance instance-00000053 to attach disk [datastore1] 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66/4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 886.074175] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bb3d9c3c-6205-456f-bedb-31c4366d97b1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.083982] env[62385]: DEBUG nova.virt.hardware [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 886.084241] env[62385]: DEBUG nova.virt.hardware [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 886.084404] env[62385]: DEBUG nova.virt.hardware [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 886.084588] env[62385]: DEBUG nova.virt.hardware [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 886.084738] env[62385]: DEBUG nova.virt.hardware [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 886.084890] env[62385]: DEBUG nova.virt.hardware [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 886.085170] env[62385]: DEBUG nova.virt.hardware [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 886.085349] env[62385]: DEBUG nova.virt.hardware [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 886.085525] env[62385]: DEBUG nova.virt.hardware [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 886.085692] env[62385]: DEBUG nova.virt.hardware [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 886.085869] env[62385]: DEBUG nova.virt.hardware [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 886.086780] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79762bae-7efe-448f-8087-42cd776258e0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.090660] env[62385]: DEBUG oslo_vmware.api [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Waiting for the task: (returnval){ [ 886.090660] env[62385]: value = "task-1206025" [ 886.090660] env[62385]: _type = "Task" [ 886.090660] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.097594] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adc9c924-657e-447a-a165-7f92c38d6175 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.104597] env[62385]: DEBUG oslo_vmware.api [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Task: {'id': task-1206025, 'name': Rename_Task} progress is 10%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.118972] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 886.119689] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:74:68:57', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ee20e439-fed9-490e-97dd-f3c886977ae1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '72d846e7-a597-4eb7-87ea-0d4309ceb42b', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 886.130849] env[62385]: DEBUG oslo.service.loopingcall [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 886.131261] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-64222340-49de-4c46-a9cd-5a298c817b9a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.133898] env[62385]: DEBUG nova.scheduler.client.report [None req-032bafda-a74f-4b9e-8041-96a5dea4e54f tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 886.137496] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 886.137794] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5d089832-605c-4cdd-acfc-87cdc0f0b58f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.160792] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 886.160792] env[62385]: value = "task-1206027" [ 886.160792] env[62385]: _type = "Task" [ 886.160792] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.162308] env[62385]: DEBUG oslo_vmware.api [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 886.162308] env[62385]: value = "task-1206026" [ 886.162308] env[62385]: _type = "Task" [ 886.162308] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.176242] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206027, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.180342] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] VM already powered off {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 886.180679] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 886.181668] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53f76934-768c-4b7a-8c16-f66dd3b2dc2c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.191155] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 886.191442] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e665c830-7da9-4493-b476-d690147434e3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.300346] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 886.300633] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Deleting contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 886.300831] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Deleting the datastore file [datastore2] d12ffea4-2275-4974-a5c6-b7fa365865b4 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 886.301571] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ca6eeade-3fec-4771-b7af-d0bcbe4519bc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.309603] env[62385]: DEBUG oslo_vmware.api [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 886.309603] env[62385]: value = "task-1206029" [ 886.309603] env[62385]: _type = "Task" [ 886.309603] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.319917] env[62385]: DEBUG oslo_vmware.api [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206029, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.406730] env[62385]: DEBUG oslo_vmware.api [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Task: {'id': task-1206024, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.601870] env[62385]: DEBUG oslo_vmware.api [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Task: {'id': task-1206025, 'name': Rename_Task, 'duration_secs': 0.139859} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.601870] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 886.602119] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-871862db-b2a5-41cb-9cec-f23c5cf277a0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.608823] env[62385]: DEBUG oslo_vmware.api [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Waiting for the task: (returnval){ [ 886.608823] env[62385]: value = "task-1206030" [ 886.608823] env[62385]: _type = "Task" [ 886.608823] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.617406] env[62385]: DEBUG oslo_vmware.api [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Task: {'id': task-1206030, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.656355] env[62385]: DEBUG oslo_concurrency.lockutils [None req-032bafda-a74f-4b9e-8041-96a5dea4e54f tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.974s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.659250] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7b5bd489-2f81-4572-8b68-daa320f73dc8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.129s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.659506] env[62385]: DEBUG nova.objects.instance [None req-7b5bd489-2f81-4572-8b68-daa320f73dc8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Lazy-loading 'resources' on Instance uuid 2a579d13-5372-4340-b7b3-cc02c1912624 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 886.672052] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206027, 'name': CreateVM_Task, 'duration_secs': 0.451072} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.672327] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 886.673058] env[62385]: DEBUG oslo_concurrency.lockutils [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.673247] env[62385]: DEBUG oslo_concurrency.lockutils [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.673468] env[62385]: DEBUG oslo_concurrency.lockutils [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 886.673754] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1fff0fb-894d-4bfa-86d5-a411d8567ce4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.676231] env[62385]: INFO nova.scheduler.client.report [None req-032bafda-a74f-4b9e-8041-96a5dea4e54f tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Deleted allocations for instance 699a57a9-5a1a-4cd1-8449-723400d9caf4 [ 886.681091] env[62385]: DEBUG oslo_vmware.api [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 886.681091] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e12e8e-a920-04df-54e3-b0a589c95ec2" [ 886.681091] env[62385]: _type = "Task" [ 886.681091] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.690374] env[62385]: DEBUG oslo_vmware.api [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e12e8e-a920-04df-54e3-b0a589c95ec2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.703880] env[62385]: DEBUG nova.compute.manager [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 886.727218] env[62385]: DEBUG nova.virt.hardware [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 886.727523] env[62385]: DEBUG nova.virt.hardware [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 886.727717] env[62385]: DEBUG nova.virt.hardware [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 886.727956] env[62385]: DEBUG nova.virt.hardware [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 886.728149] env[62385]: DEBUG nova.virt.hardware [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 886.728314] env[62385]: DEBUG nova.virt.hardware [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 886.728528] env[62385]: DEBUG nova.virt.hardware [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 886.728693] env[62385]: DEBUG nova.virt.hardware [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 886.728865] env[62385]: DEBUG nova.virt.hardware [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 886.729075] env[62385]: DEBUG nova.virt.hardware [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 886.729290] env[62385]: DEBUG nova.virt.hardware [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 886.730693] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf42e47b-92c6-4cf4-87d5-e49be5340551 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.739780] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48c34ac8-292d-47e4-ba9d-10ad9bac95af {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.754315] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Instance VIF info [] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 886.760036] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Creating folder: Project (c00d0b20367f4deea34717c1fbd65467). Parent ref: group-v261107. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 886.760624] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0dcb66f9-6d59-4619-a1dc-57dccd98dafe {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.772971] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Created folder: Project (c00d0b20367f4deea34717c1fbd65467) in parent group-v261107. [ 886.773256] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Creating folder: Instances. Parent ref: group-v261248. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 886.773480] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2428373c-5316-4952-9c66-5ef366a397f1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.783346] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Created folder: Instances in parent group-v261248. [ 886.783599] env[62385]: DEBUG oslo.service.loopingcall [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 886.783798] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 886.784050] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7e5c2810-db3b-413a-ba22-b2e548d11d17 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.802360] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 886.802360] env[62385]: value = "task-1206033" [ 886.802360] env[62385]: _type = "Task" [ 886.802360] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.810679] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206033, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.818582] env[62385]: DEBUG oslo_vmware.api [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206029, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.269096} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.818841] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 886.819956] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Deleted contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 886.819956] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 886.907372] env[62385]: DEBUG oslo_vmware.api [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Task: {'id': task-1206024, 'name': PowerOnVM_Task, 'duration_secs': 0.614097} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.907745] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 886.908036] env[62385]: INFO nova.compute.manager [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Took 9.23 seconds to spawn the instance on the hypervisor. [ 886.908289] env[62385]: DEBUG nova.compute.manager [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 886.909306] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a5a6278-3e93-4ad9-8336-3725e0376b88 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.119760] env[62385]: DEBUG oslo_vmware.api [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Task: {'id': task-1206030, 'name': PowerOnVM_Task, 'duration_secs': 0.473695} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.119963] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 887.120218] env[62385]: INFO nova.compute.manager [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Took 4.88 seconds to spawn the instance on the hypervisor. [ 887.120464] env[62385]: DEBUG nova.compute.manager [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 887.121278] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0cb0602-96f1-4d01-81ca-f309962353e8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.192727] env[62385]: DEBUG oslo_concurrency.lockutils [None req-032bafda-a74f-4b9e-8041-96a5dea4e54f tempest-ServerGroupTestJSON-1927926256 tempest-ServerGroupTestJSON-1927926256-project-member] Lock "699a57a9-5a1a-4cd1-8449-723400d9caf4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.822s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.201620] env[62385]: DEBUG oslo_vmware.api [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e12e8e-a920-04df-54e3-b0a589c95ec2, 'name': SearchDatastore_Task, 'duration_secs': 0.0248} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.202090] env[62385]: DEBUG oslo_concurrency.lockutils [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.202412] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 887.202679] env[62385]: DEBUG oslo_concurrency.lockutils [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.202893] env[62385]: DEBUG oslo_concurrency.lockutils [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.203019] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 887.203310] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-47ac387f-7d05-4119-89e3-ae35dbc888e3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.217997] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 887.218247] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 887.219281] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5a1317c1-141b-4c67-869d-2a0e5b771f36 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.229243] env[62385]: DEBUG oslo_vmware.api [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 887.229243] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521d37ee-3c4f-c7ff-fe79-b0c07b11c42c" [ 887.229243] env[62385]: _type = "Task" [ 887.229243] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.240524] env[62385]: DEBUG oslo_vmware.api [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521d37ee-3c4f-c7ff-fe79-b0c07b11c42c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.314842] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206033, 'name': CreateVM_Task, 'duration_secs': 0.456732} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.315077] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 887.315458] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.315673] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.315966] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 887.316244] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b00b0c7-f329-48cd-97a6-944cabb28959 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.321575] env[62385]: DEBUG oslo_vmware.api [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Waiting for the task: (returnval){ [ 887.321575] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5298a909-7b5c-6d9f-59a0-530170d7aefa" [ 887.321575] env[62385]: _type = "Task" [ 887.321575] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.337091] env[62385]: DEBUG oslo_vmware.api [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5298a909-7b5c-6d9f-59a0-530170d7aefa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.432556] env[62385]: INFO nova.compute.manager [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Took 33.43 seconds to build instance. [ 887.533371] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-976c132d-4227-414f-b209-d13c5ec896a9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.544520] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97b8ace5-4df1-43cc-8047-82d94e12b936 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.581194] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-726fa0c6-6455-4809-9665-65b9ccd3a367 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.589906] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3671f94-6768-4cbb-a1d7-95cf8cda11a2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.605446] env[62385]: DEBUG nova.compute.provider_tree [None req-7b5bd489-2f81-4572-8b68-daa320f73dc8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 887.638884] env[62385]: INFO nova.compute.manager [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Took 31.03 seconds to build instance. [ 887.741595] env[62385]: DEBUG oslo_vmware.api [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521d37ee-3c4f-c7ff-fe79-b0c07b11c42c, 'name': SearchDatastore_Task, 'duration_secs': 0.018392} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.742428] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be6b90e7-ec9b-4845-9144-d9a24fa31117 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.749534] env[62385]: DEBUG oslo_vmware.api [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 887.749534] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5214eb2a-6dc7-ec76-c9e6-21dda469f62c" [ 887.749534] env[62385]: _type = "Task" [ 887.749534] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.761424] env[62385]: DEBUG oslo_vmware.api [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5214eb2a-6dc7-ec76-c9e6-21dda469f62c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.835595] env[62385]: DEBUG oslo_vmware.api [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5298a909-7b5c-6d9f-59a0-530170d7aefa, 'name': SearchDatastore_Task, 'duration_secs': 0.017462} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.836101] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.836478] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 887.836701] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.900796] env[62385]: DEBUG nova.virt.hardware [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 887.900796] env[62385]: DEBUG nova.virt.hardware [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 887.900796] env[62385]: DEBUG nova.virt.hardware [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 887.900983] env[62385]: DEBUG nova.virt.hardware [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 887.901135] env[62385]: DEBUG nova.virt.hardware [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 887.901283] env[62385]: DEBUG nova.virt.hardware [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 887.901493] env[62385]: DEBUG nova.virt.hardware [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 887.901768] env[62385]: DEBUG nova.virt.hardware [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 887.902820] env[62385]: DEBUG nova.virt.hardware [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 887.902820] env[62385]: DEBUG nova.virt.hardware [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 887.902820] env[62385]: DEBUG nova.virt.hardware [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 887.904068] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0482868-71df-439d-8364-66e84bc2dee0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.914206] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-591b4506-4ce5-4769-b438-0532a4afead2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.931388] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:80:c9:48', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4adc8ed0-d11a-4510-9be0-b27c0da3a903', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0e719241-7a39-42dd-b37e-31cc75a4da74', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 887.940040] env[62385]: DEBUG oslo.service.loopingcall [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 887.941395] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8177566d-692c-4e2a-a0d1-a38b364c6127 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Lock "e9bf98d1-089a-4078-b2f8-eb77943731e5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.402s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.941395] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 887.941395] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-71c60a75-dc51-4b61-9365-0a7d0c60bb9c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.962568] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 887.962568] env[62385]: value = "task-1206034" [ 887.962568] env[62385]: _type = "Task" [ 887.962568] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.972151] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206034, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.112263] env[62385]: DEBUG nova.scheduler.client.report [None req-7b5bd489-2f81-4572-8b68-daa320f73dc8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 888.144470] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3e42bf0f-b01a-4236-9822-a15e0604e110 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Lock "4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.765s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.261611] env[62385]: DEBUG oslo_vmware.api [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5214eb2a-6dc7-ec76-c9e6-21dda469f62c, 'name': SearchDatastore_Task, 'duration_secs': 0.034356} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.264062] env[62385]: DEBUG oslo_concurrency.lockutils [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.264062] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 33387505-c576-488b-8c9c-b064fe81a7d7/33387505-c576-488b-8c9c-b064fe81a7d7.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 888.264062] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.264062] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 888.264062] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d03cc008-ccf5-4cb8-896f-5bdbc4e3f982 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.266548] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6db1f5a7-3a93-4dc3-b558-5fb351b7a6c4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.276499] env[62385]: DEBUG oslo_vmware.api [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 888.276499] env[62385]: value = "task-1206035" [ 888.276499] env[62385]: _type = "Task" [ 888.276499] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.282714] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 888.282714] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 888.284425] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e89d6d99-241a-41da-b644-fc6d84e505cd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.290966] env[62385]: DEBUG oslo_vmware.api [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206035, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.295318] env[62385]: DEBUG oslo_vmware.api [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Waiting for the task: (returnval){ [ 888.295318] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5265e197-2361-262f-cc78-d34246aef9da" [ 888.295318] env[62385]: _type = "Task" [ 888.295318] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.304616] env[62385]: DEBUG oslo_vmware.api [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5265e197-2361-262f-cc78-d34246aef9da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.398552] env[62385]: DEBUG nova.compute.manager [None req-a63de1c8-d0e9-4063-b934-8b9bfb53db2d tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 888.400264] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c88c405-3965-476d-8d85-e73e6655c956 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.474798] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206034, 'name': CreateVM_Task, 'duration_secs': 0.375314} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.475288] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 888.475837] env[62385]: DEBUG oslo_concurrency.lockutils [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.476029] env[62385]: DEBUG oslo_concurrency.lockutils [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.476382] env[62385]: DEBUG oslo_concurrency.lockutils [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 888.476708] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b1f3d58f-a9b4-4af4-b359-5bad3081ed48 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.482427] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0f12566d-f7e2-4d77-b350-830290273675 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Acquiring lock "4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.482427] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0f12566d-f7e2-4d77-b350-830290273675 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Lock "4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.482427] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0f12566d-f7e2-4d77-b350-830290273675 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Acquiring lock "4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.482427] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0f12566d-f7e2-4d77-b350-830290273675 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Lock "4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.482427] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0f12566d-f7e2-4d77-b350-830290273675 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Lock "4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.486605] env[62385]: DEBUG oslo_vmware.api [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 888.486605] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]522f2456-b4b7-5492-5460-63f6766ce103" [ 888.486605] env[62385]: _type = "Task" [ 888.486605] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.486828] env[62385]: INFO nova.compute.manager [None req-0f12566d-f7e2-4d77-b350-830290273675 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Terminating instance [ 888.494773] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0f12566d-f7e2-4d77-b350-830290273675 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Acquiring lock "refresh_cache-4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.494773] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0f12566d-f7e2-4d77-b350-830290273675 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Acquired lock "refresh_cache-4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.494773] env[62385]: DEBUG nova.network.neutron [None req-0f12566d-f7e2-4d77-b350-830290273675 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 888.502074] env[62385]: DEBUG oslo_vmware.api [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]522f2456-b4b7-5492-5460-63f6766ce103, 'name': SearchDatastore_Task, 'duration_secs': 0.012432} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.502074] env[62385]: DEBUG oslo_concurrency.lockutils [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.502074] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 888.502074] env[62385]: DEBUG oslo_concurrency.lockutils [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.618534] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7b5bd489-2f81-4572-8b68-daa320f73dc8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.959s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.621400] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.938s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.623966] env[62385]: INFO nova.compute.claims [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 888.658109] env[62385]: INFO nova.scheduler.client.report [None req-7b5bd489-2f81-4572-8b68-daa320f73dc8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Deleted allocations for instance 2a579d13-5372-4340-b7b3-cc02c1912624 [ 888.788395] env[62385]: DEBUG oslo_vmware.api [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206035, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.808916] env[62385]: DEBUG oslo_vmware.api [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5265e197-2361-262f-cc78-d34246aef9da, 'name': SearchDatastore_Task, 'duration_secs': 0.012503} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.809913] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c7b358b9-73b3-4de5-80ae-397585457244 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.816750] env[62385]: DEBUG oslo_vmware.api [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Waiting for the task: (returnval){ [ 888.816750] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52cb86b4-7be7-63ae-026e-9af7f0ffae89" [ 888.816750] env[62385]: _type = "Task" [ 888.816750] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.829136] env[62385]: DEBUG oslo_vmware.api [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52cb86b4-7be7-63ae-026e-9af7f0ffae89, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.915022] env[62385]: INFO nova.compute.manager [None req-a63de1c8-d0e9-4063-b934-8b9bfb53db2d tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] instance snapshotting [ 888.916023] env[62385]: DEBUG nova.objects.instance [None req-a63de1c8-d0e9-4063-b934-8b9bfb53db2d tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Lazy-loading 'flavor' on Instance uuid 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 889.014661] env[62385]: DEBUG nova.network.neutron [None req-0f12566d-f7e2-4d77-b350-830290273675 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 889.090805] env[62385]: DEBUG nova.network.neutron [None req-0f12566d-f7e2-4d77-b350-830290273675 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.170692] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7b5bd489-2f81-4572-8b68-daa320f73dc8 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Lock "2a579d13-5372-4340-b7b3-cc02c1912624" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.327s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.288068] env[62385]: DEBUG oslo_vmware.api [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206035, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.600108} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.288423] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 33387505-c576-488b-8c9c-b064fe81a7d7/33387505-c576-488b-8c9c-b064fe81a7d7.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 889.288647] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 889.288914] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-22f4667c-a227-4880-aa09-69adbedbe3b9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.297681] env[62385]: DEBUG oslo_vmware.api [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 889.297681] env[62385]: value = "task-1206036" [ 889.297681] env[62385]: _type = "Task" [ 889.297681] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.309359] env[62385]: DEBUG oslo_vmware.api [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206036, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.329943] env[62385]: DEBUG oslo_vmware.api [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52cb86b4-7be7-63ae-026e-9af7f0ffae89, 'name': SearchDatastore_Task, 'duration_secs': 0.039447} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.330287] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.330567] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 2bcd9457-8a73-4e7a-b778-d52c468b3aae/2bcd9457-8a73-4e7a-b778-d52c468b3aae.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 889.330903] env[62385]: DEBUG oslo_concurrency.lockutils [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.331109] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 889.331376] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2d077994-9778-4bcb-bbed-99b6a98a1f0a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.333810] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e0586a29-ab24-40b9-91d8-ea97737f493c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.341567] env[62385]: DEBUG oslo_vmware.api [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Waiting for the task: (returnval){ [ 889.341567] env[62385]: value = "task-1206037" [ 889.341567] env[62385]: _type = "Task" [ 889.341567] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.345962] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 889.346324] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 889.347321] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12a87bd8-b80c-4463-ae0d-b373df8bec98 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.353559] env[62385]: DEBUG oslo_vmware.api [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': task-1206037, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.355043] env[62385]: DEBUG oslo_vmware.api [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 889.355043] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5218b708-929e-06e1-5139-ce0c8aab6aeb" [ 889.355043] env[62385]: _type = "Task" [ 889.355043] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.364166] env[62385]: DEBUG oslo_vmware.api [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5218b708-929e-06e1-5139-ce0c8aab6aeb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.422071] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71c36300-2ffb-41d7-876d-7698bdf597b8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.441071] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c55b200c-6f81-490c-9ee4-dd6ed4996ac1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.595068] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0f12566d-f7e2-4d77-b350-830290273675 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Releasing lock "refresh_cache-4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.595322] env[62385]: DEBUG nova.compute.manager [None req-0f12566d-f7e2-4d77-b350-830290273675 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 889.595978] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0f12566d-f7e2-4d77-b350-830290273675 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 889.596713] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aab618aa-d5ed-49b8-a444-80bfa2faac4d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.609291] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f12566d-f7e2-4d77-b350-830290273675 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 889.609291] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-799f77d0-f9c5-4a15-9027-395fb1a2a8b7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.617689] env[62385]: DEBUG oslo_vmware.api [None req-0f12566d-f7e2-4d77-b350-830290273675 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Waiting for the task: (returnval){ [ 889.617689] env[62385]: value = "task-1206038" [ 889.617689] env[62385]: _type = "Task" [ 889.617689] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.628581] env[62385]: DEBUG oslo_vmware.api [None req-0f12566d-f7e2-4d77-b350-830290273675 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Task: {'id': task-1206038, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.810221] env[62385]: DEBUG oslo_vmware.api [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206036, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.113604} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.810557] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 889.812044] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d50f62cf-a09a-4287-99cf-b001032abe9c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.839270] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Reconfiguring VM instance instance-0000002c to attach disk [datastore2] 33387505-c576-488b-8c9c-b064fe81a7d7/33387505-c576-488b-8c9c-b064fe81a7d7.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 889.842477] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b9982210-0399-4f35-a5e5-1a21c95faeb1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.869784] env[62385]: DEBUG oslo_vmware.api [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': task-1206037, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.877533] env[62385]: DEBUG oslo_vmware.api [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5218b708-929e-06e1-5139-ce0c8aab6aeb, 'name': SearchDatastore_Task, 'duration_secs': 0.021908} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.878085] env[62385]: DEBUG oslo_vmware.api [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 889.878085] env[62385]: value = "task-1206039" [ 889.878085] env[62385]: _type = "Task" [ 889.878085] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.879055] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fbac39a8-56f7-47fc-afef-e7251b622c86 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.891273] env[62385]: DEBUG oslo_vmware.api [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 889.891273] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52070025-d622-bd63-5cb3-733e0ad4269b" [ 889.891273] env[62385]: _type = "Task" [ 889.891273] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.897683] env[62385]: DEBUG oslo_vmware.api [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206039, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.907062] env[62385]: DEBUG oslo_vmware.api [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52070025-d622-bd63-5cb3-733e0ad4269b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.953220] env[62385]: DEBUG nova.compute.manager [None req-a63de1c8-d0e9-4063-b934-8b9bfb53db2d tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Instance disappeared during snapshot {{(pid=62385) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 890.001341] env[62385]: DEBUG oslo_concurrency.lockutils [None req-620c0e7e-c879-4917-9593-e5c8b9ecfde9 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Acquiring lock "e9bf98d1-089a-4078-b2f8-eb77943731e5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.002334] env[62385]: DEBUG oslo_concurrency.lockutils [None req-620c0e7e-c879-4917-9593-e5c8b9ecfde9 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Lock "e9bf98d1-089a-4078-b2f8-eb77943731e5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.002661] env[62385]: DEBUG oslo_concurrency.lockutils [None req-620c0e7e-c879-4917-9593-e5c8b9ecfde9 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Acquiring lock "e9bf98d1-089a-4078-b2f8-eb77943731e5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.002943] env[62385]: DEBUG oslo_concurrency.lockutils [None req-620c0e7e-c879-4917-9593-e5c8b9ecfde9 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Lock "e9bf98d1-089a-4078-b2f8-eb77943731e5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.003226] env[62385]: DEBUG oslo_concurrency.lockutils [None req-620c0e7e-c879-4917-9593-e5c8b9ecfde9 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Lock "e9bf98d1-089a-4078-b2f8-eb77943731e5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.006658] env[62385]: INFO nova.compute.manager [None req-620c0e7e-c879-4917-9593-e5c8b9ecfde9 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Terminating instance [ 890.011255] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73db0f25-9985-4f98-aa23-2e761f2ceda6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.016250] env[62385]: DEBUG nova.compute.manager [None req-620c0e7e-c879-4917-9593-e5c8b9ecfde9 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 890.016580] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-620c0e7e-c879-4917-9593-e5c8b9ecfde9 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 890.017712] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac5ee896-1c68-41eb-a260-78df0124e4f1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.034091] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05833fcb-06bb-443a-bcc5-741149f0ab70 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.039496] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-620c0e7e-c879-4917-9593-e5c8b9ecfde9 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 890.039813] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-698a24b0-38de-4e29-a5e3-cec7dbaeee42 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.075829] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a79ac3e-e444-44fe-a481-555ef3882c27 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.078305] env[62385]: DEBUG oslo_vmware.api [None req-620c0e7e-c879-4917-9593-e5c8b9ecfde9 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Waiting for the task: (returnval){ [ 890.078305] env[62385]: value = "task-1206040" [ 890.078305] env[62385]: _type = "Task" [ 890.078305] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.086462] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47915192-22b4-45cb-9b4e-5bd5e2874575 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.094451] env[62385]: DEBUG oslo_vmware.api [None req-620c0e7e-c879-4917-9593-e5c8b9ecfde9 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Task: {'id': task-1206040, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.106601] env[62385]: DEBUG nova.compute.provider_tree [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 890.131105] env[62385]: DEBUG oslo_vmware.api [None req-0f12566d-f7e2-4d77-b350-830290273675 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Task: {'id': task-1206038, 'name': PowerOffVM_Task, 'duration_secs': 0.189251} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.131586] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f12566d-f7e2-4d77-b350-830290273675 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 890.131804] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0f12566d-f7e2-4d77-b350-830290273675 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 890.133232] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-da8d4e42-0ebd-403e-9a61-f002e69f5532 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.152938] env[62385]: DEBUG nova.compute.manager [None req-a63de1c8-d0e9-4063-b934-8b9bfb53db2d tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Found 0 images (rotation: 2) {{(pid=62385) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 890.165048] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0f12566d-f7e2-4d77-b350-830290273675 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 890.165702] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0f12566d-f7e2-4d77-b350-830290273675 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 890.165803] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f12566d-f7e2-4d77-b350-830290273675 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Deleting the datastore file [datastore1] 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 890.166051] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a14ac87d-b011-4647-b14c-2d4dd78cc777 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.174162] env[62385]: DEBUG oslo_vmware.api [None req-0f12566d-f7e2-4d77-b350-830290273675 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Waiting for the task: (returnval){ [ 890.174162] env[62385]: value = "task-1206042" [ 890.174162] env[62385]: _type = "Task" [ 890.174162] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.184654] env[62385]: DEBUG oslo_vmware.api [None req-0f12566d-f7e2-4d77-b350-830290273675 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Task: {'id': task-1206042, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.230125] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Acquiring lock "09c017a9-ad18-49b6-b72d-958023c81b24" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.230478] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Lock "09c017a9-ad18-49b6-b72d-958023c81b24" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.371831] env[62385]: DEBUG oslo_vmware.api [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': task-1206037, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.709596} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.372287] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 2bcd9457-8a73-4e7a-b778-d52c468b3aae/2bcd9457-8a73-4e7a-b778-d52c468b3aae.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 890.372625] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 890.373016] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3ce1c396-fa4f-4148-b53c-af2b196b2342 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.386242] env[62385]: DEBUG oslo_vmware.api [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Waiting for the task: (returnval){ [ 890.386242] env[62385]: value = "task-1206043" [ 890.386242] env[62385]: _type = "Task" [ 890.386242] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.395970] env[62385]: DEBUG oslo_vmware.api [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206039, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.405845] env[62385]: DEBUG oslo_vmware.api [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': task-1206043, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.412725] env[62385]: DEBUG oslo_vmware.api [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52070025-d622-bd63-5cb3-733e0ad4269b, 'name': SearchDatastore_Task, 'duration_secs': 0.060578} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.413048] env[62385]: DEBUG oslo_concurrency.lockutils [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.413340] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] d12ffea4-2275-4974-a5c6-b7fa365865b4/d12ffea4-2275-4974-a5c6-b7fa365865b4.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 890.413637] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-db9defb2-9432-4e99-8afc-408b8342450c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.421303] env[62385]: DEBUG oslo_vmware.api [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 890.421303] env[62385]: value = "task-1206044" [ 890.421303] env[62385]: _type = "Task" [ 890.421303] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.432274] env[62385]: DEBUG oslo_vmware.api [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206044, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.590086] env[62385]: DEBUG oslo_vmware.api [None req-620c0e7e-c879-4917-9593-e5c8b9ecfde9 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Task: {'id': task-1206040, 'name': PowerOffVM_Task, 'duration_secs': 0.218334} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.590555] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-620c0e7e-c879-4917-9593-e5c8b9ecfde9 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 890.590850] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-620c0e7e-c879-4917-9593-e5c8b9ecfde9 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 890.591239] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c916c360-c6ce-4ae1-9118-3aac039d4bc0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.613214] env[62385]: DEBUG nova.scheduler.client.report [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 890.671735] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-620c0e7e-c879-4917-9593-e5c8b9ecfde9 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 890.671735] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-620c0e7e-c879-4917-9593-e5c8b9ecfde9 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 890.671919] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-620c0e7e-c879-4917-9593-e5c8b9ecfde9 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Deleting the datastore file [datastore1] e9bf98d1-089a-4078-b2f8-eb77943731e5 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 890.672250] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-772d77c5-e740-4b87-a625-dfa27d1c3d95 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.680278] env[62385]: DEBUG oslo_vmware.api [None req-620c0e7e-c879-4917-9593-e5c8b9ecfde9 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Waiting for the task: (returnval){ [ 890.680278] env[62385]: value = "task-1206046" [ 890.680278] env[62385]: _type = "Task" [ 890.680278] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.686507] env[62385]: DEBUG oslo_vmware.api [None req-0f12566d-f7e2-4d77-b350-830290273675 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Task: {'id': task-1206042, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.16655} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.686804] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f12566d-f7e2-4d77-b350-830290273675 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 890.686979] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0f12566d-f7e2-4d77-b350-830290273675 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 890.687148] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0f12566d-f7e2-4d77-b350-830290273675 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 890.687326] env[62385]: INFO nova.compute.manager [None req-0f12566d-f7e2-4d77-b350-830290273675 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Took 1.09 seconds to destroy the instance on the hypervisor. [ 890.687576] env[62385]: DEBUG oslo.service.loopingcall [None req-0f12566d-f7e2-4d77-b350-830290273675 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 890.688175] env[62385]: DEBUG nova.compute.manager [-] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 890.688337] env[62385]: DEBUG nova.network.neutron [-] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 890.693374] env[62385]: DEBUG oslo_vmware.api [None req-620c0e7e-c879-4917-9593-e5c8b9ecfde9 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Task: {'id': task-1206046, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.708090] env[62385]: DEBUG nova.network.neutron [-] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 890.733340] env[62385]: DEBUG nova.compute.manager [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 890.891546] env[62385]: DEBUG oslo_vmware.api [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206039, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.900155] env[62385]: DEBUG oslo_vmware.api [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': task-1206043, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080548} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.900815] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 890.901684] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d83c966-fd3c-487b-94d4-0c2a7e1676e6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.922884] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] 2bcd9457-8a73-4e7a-b778-d52c468b3aae/2bcd9457-8a73-4e7a-b778-d52c468b3aae.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 890.923278] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cf9d7b2f-0f91-4f44-b49f-8d6813f74e64 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.948054] env[62385]: DEBUG oslo_vmware.api [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206044, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.949571] env[62385]: DEBUG oslo_vmware.api [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Waiting for the task: (returnval){ [ 890.949571] env[62385]: value = "task-1206047" [ 890.949571] env[62385]: _type = "Task" [ 890.949571] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.958067] env[62385]: DEBUG oslo_vmware.api [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': task-1206047, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.118109] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.497s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.118739] env[62385]: DEBUG nova.compute.manager [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 891.121783] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a61cdaf1-3673-4b1a-b5bb-58fa8fda5a47 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.434s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.122043] env[62385]: DEBUG nova.objects.instance [None req-a61cdaf1-3673-4b1a-b5bb-58fa8fda5a47 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lazy-loading 'resources' on Instance uuid fe0624a9-09a8-498b-bb3c-fda6cab92341 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 891.193588] env[62385]: DEBUG oslo_vmware.api [None req-620c0e7e-c879-4917-9593-e5c8b9ecfde9 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Task: {'id': task-1206046, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.205841} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.193588] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-620c0e7e-c879-4917-9593-e5c8b9ecfde9 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 891.193588] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-620c0e7e-c879-4917-9593-e5c8b9ecfde9 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 891.193893] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-620c0e7e-c879-4917-9593-e5c8b9ecfde9 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 891.193893] env[62385]: INFO nova.compute.manager [None req-620c0e7e-c879-4917-9593-e5c8b9ecfde9 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Took 1.18 seconds to destroy the instance on the hypervisor. [ 891.194299] env[62385]: DEBUG oslo.service.loopingcall [None req-620c0e7e-c879-4917-9593-e5c8b9ecfde9 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 891.194523] env[62385]: DEBUG nova.compute.manager [-] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 891.194621] env[62385]: DEBUG nova.network.neutron [-] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 891.211163] env[62385]: DEBUG nova.network.neutron [-] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.257232] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.391556] env[62385]: DEBUG oslo_vmware.api [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206039, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.434765] env[62385]: DEBUG oslo_vmware.api [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206044, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.461537] env[62385]: DEBUG oslo_vmware.api [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': task-1206047, 'name': ReconfigVM_Task, 'duration_secs': 0.331845} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.461537] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Reconfigured VM instance instance-00000054 to attach disk [datastore2] 2bcd9457-8a73-4e7a-b778-d52c468b3aae/2bcd9457-8a73-4e7a-b778-d52c468b3aae.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 891.461537] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5dbcbc7a-a31a-4c7f-a01f-b4a93a1afec2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.470913] env[62385]: DEBUG oslo_vmware.api [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Waiting for the task: (returnval){ [ 891.470913] env[62385]: value = "task-1206048" [ 891.470913] env[62385]: _type = "Task" [ 891.470913] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.479980] env[62385]: DEBUG oslo_vmware.api [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': task-1206048, 'name': Rename_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.495569] env[62385]: DEBUG nova.compute.manager [req-6fd949ca-6e4f-41d7-a085-e22c549fa845 req-1d3e4883-94d1-4504-938f-e4048efe606a service nova] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Received event network-vif-deleted-feaac30d-d553-4422-8c98-7a74ee001276 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 891.495698] env[62385]: INFO nova.compute.manager [req-6fd949ca-6e4f-41d7-a085-e22c549fa845 req-1d3e4883-94d1-4504-938f-e4048efe606a service nova] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Neutron deleted interface feaac30d-d553-4422-8c98-7a74ee001276; detaching it from the instance and deleting it from the info cache [ 891.496348] env[62385]: DEBUG nova.network.neutron [req-6fd949ca-6e4f-41d7-a085-e22c549fa845 req-1d3e4883-94d1-4504-938f-e4048efe606a service nova] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.610918] env[62385]: DEBUG oslo_vmware.rw_handles [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528864e4-6de8-fe59-acb2-32151aebccca/disk-0.vmdk. {{(pid=62385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 891.612009] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f604470c-00a7-4b54-b06c-033227118d6c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.621682] env[62385]: DEBUG oslo_vmware.rw_handles [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528864e4-6de8-fe59-acb2-32151aebccca/disk-0.vmdk is in state: ready. {{(pid=62385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 891.621890] env[62385]: ERROR oslo_vmware.rw_handles [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528864e4-6de8-fe59-acb2-32151aebccca/disk-0.vmdk due to incomplete transfer. [ 891.622303] env[62385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-814db8fc-22f0-4498-a022-78cbe385f8ad {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.625493] env[62385]: DEBUG nova.compute.utils [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 891.631359] env[62385]: DEBUG nova.compute.manager [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 891.631619] env[62385]: DEBUG nova.network.neutron [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 891.643055] env[62385]: DEBUG oslo_vmware.rw_handles [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528864e4-6de8-fe59-acb2-32151aebccca/disk-0.vmdk. {{(pid=62385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 891.643055] env[62385]: DEBUG nova.virt.vmwareapi.images [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Uploaded image aa5b9007-7bfe-4917-a311-121e4cac3174 to the Glance image server {{(pid=62385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 891.645376] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Destroying the VM {{(pid=62385) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 891.646282] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-eb1399c0-53cf-474f-b41d-caa870d17196 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.657173] env[62385]: DEBUG oslo_vmware.api [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 891.657173] env[62385]: value = "task-1206049" [ 891.657173] env[62385]: _type = "Task" [ 891.657173] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.670463] env[62385]: DEBUG oslo_vmware.api [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206049, 'name': Destroy_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.677084] env[62385]: DEBUG nova.policy [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5040b20c848c4d28a198f1773a3ff21e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '012426f098ce40c3aaa00f628fe9cebb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 891.713571] env[62385]: INFO nova.compute.manager [-] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Took 1.03 seconds to deallocate network for instance. [ 891.890811] env[62385]: DEBUG oslo_vmware.api [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206039, 'name': ReconfigVM_Task, 'duration_secs': 1.593929} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.894084] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Reconfigured VM instance instance-0000002c to attach disk [datastore2] 33387505-c576-488b-8c9c-b064fe81a7d7/33387505-c576-488b-8c9c-b064fe81a7d7.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 891.895973] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f51b66f7-9c9f-42cb-ae07-c1ba71dac66a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.904507] env[62385]: DEBUG oslo_vmware.api [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 891.904507] env[62385]: value = "task-1206050" [ 891.904507] env[62385]: _type = "Task" [ 891.904507] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.913947] env[62385]: DEBUG oslo_vmware.api [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206050, 'name': Rename_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.935296] env[62385]: DEBUG oslo_vmware.api [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206044, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.436632} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.938163] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] d12ffea4-2275-4974-a5c6-b7fa365865b4/d12ffea4-2275-4974-a5c6-b7fa365865b4.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 891.938416] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 891.938906] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d579dfdd-02de-4198-8767-20e1be33d007 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.946280] env[62385]: DEBUG oslo_vmware.api [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 891.946280] env[62385]: value = "task-1206051" [ 891.946280] env[62385]: _type = "Task" [ 891.946280] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.960063] env[62385]: DEBUG oslo_vmware.api [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206051, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.972042] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21a19f19-8cdf-4fa0-b379-1801ebe11d0f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.977733] env[62385]: DEBUG nova.network.neutron [-] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.986311] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41da120b-f199-4e3c-97a7-f38133907668 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.990614] env[62385]: DEBUG oslo_vmware.api [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': task-1206048, 'name': Rename_Task, 'duration_secs': 0.151443} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.990752] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 891.991862] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ed5ad957-6deb-4be5-b98a-6d648975051c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.027919] env[62385]: DEBUG nova.network.neutron [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Successfully created port: 2b05b2b6-d4ce-4eb7-9d65-730d4dbad51c {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 892.030030] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-acbb4006-065c-4a9c-9619-685bb1d0e5d0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.032291] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2afc6672-8f39-420b-a346-47c4b36008a7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.035660] env[62385]: DEBUG oslo_vmware.api [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Waiting for the task: (returnval){ [ 892.035660] env[62385]: value = "task-1206052" [ 892.035660] env[62385]: _type = "Task" [ 892.035660] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.046525] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1e01e16-0d68-4ecf-9783-534178307b75 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.057416] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99310541-b1d6-4414-8147-d934ea94b9ab {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.069483] env[62385]: DEBUG oslo_vmware.api [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': task-1206052, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.080371] env[62385]: DEBUG nova.compute.provider_tree [None req-a61cdaf1-3673-4b1a-b5bb-58fa8fda5a47 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 892.093614] env[62385]: DEBUG nova.compute.manager [req-6fd949ca-6e4f-41d7-a085-e22c549fa845 req-1d3e4883-94d1-4504-938f-e4048efe606a service nova] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Detach interface failed, port_id=feaac30d-d553-4422-8c98-7a74ee001276, reason: Instance e9bf98d1-089a-4078-b2f8-eb77943731e5 could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 892.094859] env[62385]: DEBUG nova.scheduler.client.report [None req-a61cdaf1-3673-4b1a-b5bb-58fa8fda5a47 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 892.130055] env[62385]: DEBUG nova.compute.manager [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 892.168185] env[62385]: DEBUG oslo_vmware.api [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206049, 'name': Destroy_Task, 'duration_secs': 0.506967} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.168488] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Destroyed the VM [ 892.168728] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Deleting Snapshot of the VM instance {{(pid=62385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 892.169397] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-58bb6869-95ed-462d-af9a-17ee6cc28fdd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.178511] env[62385]: DEBUG oslo_vmware.api [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 892.178511] env[62385]: value = "task-1206053" [ 892.178511] env[62385]: _type = "Task" [ 892.178511] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.188060] env[62385]: DEBUG oslo_vmware.api [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206053, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.222810] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0f12566d-f7e2-4d77-b350-830290273675 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.418977] env[62385]: DEBUG oslo_vmware.api [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206050, 'name': Rename_Task, 'duration_secs': 0.193645} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.419543] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 892.419984] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-43bb8649-018d-4a2f-8af5-ad916dfebc4e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.429534] env[62385]: DEBUG oslo_vmware.api [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 892.429534] env[62385]: value = "task-1206054" [ 892.429534] env[62385]: _type = "Task" [ 892.429534] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.442816] env[62385]: DEBUG oslo_vmware.api [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206054, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.459546] env[62385]: DEBUG oslo_vmware.api [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206051, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066722} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.459833] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 892.460665] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9f3d632-7eb9-4949-ae60-ccac11d79974 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.483258] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] d12ffea4-2275-4974-a5c6-b7fa365865b4/d12ffea4-2275-4974-a5c6-b7fa365865b4.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 892.483732] env[62385]: INFO nova.compute.manager [-] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Took 1.29 seconds to deallocate network for instance. [ 892.483971] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-643ea0ce-02a6-472d-8648-b66daa79a687 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.507595] env[62385]: DEBUG oslo_vmware.api [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 892.507595] env[62385]: value = "task-1206055" [ 892.507595] env[62385]: _type = "Task" [ 892.507595] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.516930] env[62385]: DEBUG oslo_vmware.api [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206055, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.548152] env[62385]: DEBUG oslo_vmware.api [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': task-1206052, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.601335] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a61cdaf1-3673-4b1a-b5bb-58fa8fda5a47 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.479s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.604173] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6a258a04-6455-49dc-9069-3e3c3f39df66 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.363s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.604173] env[62385]: DEBUG nova.objects.instance [None req-6a258a04-6455-49dc-9069-3e3c3f39df66 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Lazy-loading 'resources' on Instance uuid 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 892.628068] env[62385]: INFO nova.scheduler.client.report [None req-a61cdaf1-3673-4b1a-b5bb-58fa8fda5a47 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Deleted allocations for instance fe0624a9-09a8-498b-bb3c-fda6cab92341 [ 892.689351] env[62385]: DEBUG oslo_vmware.api [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206053, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.941842] env[62385]: DEBUG oslo_vmware.api [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206054, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.007768] env[62385]: DEBUG oslo_concurrency.lockutils [None req-620c0e7e-c879-4917-9593-e5c8b9ecfde9 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.017716] env[62385]: DEBUG oslo_vmware.api [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206055, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.047196] env[62385]: DEBUG oslo_vmware.api [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': task-1206052, 'name': PowerOnVM_Task, 'duration_secs': 0.667983} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.047333] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 893.047503] env[62385]: INFO nova.compute.manager [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Took 6.34 seconds to spawn the instance on the hypervisor. [ 893.047696] env[62385]: DEBUG nova.compute.manager [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 893.048489] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62a1d0bd-83be-49c6-b340-65fa72990818 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.137580] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a61cdaf1-3673-4b1a-b5bb-58fa8fda5a47 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "fe0624a9-09a8-498b-bb3c-fda6cab92341" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.873s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.143983] env[62385]: DEBUG nova.compute.manager [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 893.178960] env[62385]: DEBUG nova.virt.hardware [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 893.179288] env[62385]: DEBUG nova.virt.hardware [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 893.179465] env[62385]: DEBUG nova.virt.hardware [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 893.179961] env[62385]: DEBUG nova.virt.hardware [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 893.179961] env[62385]: DEBUG nova.virt.hardware [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 893.180287] env[62385]: DEBUG nova.virt.hardware [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 893.180477] env[62385]: DEBUG nova.virt.hardware [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 893.180673] env[62385]: DEBUG nova.virt.hardware [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 893.180862] env[62385]: DEBUG nova.virt.hardware [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 893.181075] env[62385]: DEBUG nova.virt.hardware [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 893.181289] env[62385]: DEBUG nova.virt.hardware [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 893.183434] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab97da87-9691-4ab5-8488-f41fa2f0f52f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.198934] env[62385]: DEBUG oslo_vmware.api [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206053, 'name': RemoveSnapshot_Task, 'duration_secs': 0.683322} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.200510] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e60dfaa2-3d11-429e-a7e8-a38586a56a12 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.204977] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Deleted Snapshot of the VM instance {{(pid=62385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 893.205254] env[62385]: INFO nova.compute.manager [None req-26c72282-7cd7-43b0-ae06-71cf92a2988e tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Took 16.94 seconds to snapshot the instance on the hypervisor. [ 893.436801] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c4772ba-8b24-437c-835d-fab7ec78a8ec {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.444529] env[62385]: DEBUG oslo_vmware.api [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206054, 'name': PowerOnVM_Task, 'duration_secs': 0.545924} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.446363] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 893.446583] env[62385]: DEBUG nova.compute.manager [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 893.447475] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20bbe2b0-f0c8-4293-b7be-a1f216b7d7c1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.450628] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-512f1cf6-fe7f-47b4-ab95-0dcb53557a39 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.483724] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-824e8028-21f2-46f6-a561-6e845ea596ef {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.492035] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71b5178f-3290-400a-b4ca-08a684329008 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.508076] env[62385]: DEBUG nova.compute.provider_tree [None req-6a258a04-6455-49dc-9069-3e3c3f39df66 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 893.518615] env[62385]: DEBUG oslo_vmware.api [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206055, 'name': ReconfigVM_Task, 'duration_secs': 0.706557} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.519652] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Reconfigured VM instance instance-00000050 to attach disk [datastore2] d12ffea4-2275-4974-a5c6-b7fa365865b4/d12ffea4-2275-4974-a5c6-b7fa365865b4.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 893.520318] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-084760a1-af56-4ca7-92de-f813749eac69 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.527046] env[62385]: DEBUG oslo_vmware.api [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 893.527046] env[62385]: value = "task-1206056" [ 893.527046] env[62385]: _type = "Task" [ 893.527046] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.536654] env[62385]: DEBUG oslo_vmware.api [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206056, 'name': Rename_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.569336] env[62385]: INFO nova.compute.manager [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Took 34.80 seconds to build instance. [ 893.806036] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquiring lock "caab700b-bd3c-497e-b40a-eb9d20dfc1e2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.806294] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "caab700b-bd3c-497e-b40a-eb9d20dfc1e2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.834349] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "fad3b054-cf2a-4ee3-a2a5-918a96a4a125" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.834673] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "fad3b054-cf2a-4ee3-a2a5-918a96a4a125" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.981650] env[62385]: DEBUG nova.compute.manager [req-2a749a3a-4c7a-4b6e-8fdd-18c386882f72 req-ac4b40b9-1475-44d0-84ca-5d03e59407c0 service nova] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Received event network-vif-plugged-2b05b2b6-d4ce-4eb7-9d65-730d4dbad51c {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 893.982372] env[62385]: DEBUG oslo_concurrency.lockutils [req-2a749a3a-4c7a-4b6e-8fdd-18c386882f72 req-ac4b40b9-1475-44d0-84ca-5d03e59407c0 service nova] Acquiring lock "f937ef7b-ab39-4d8a-9577-7faabc652aa5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.982372] env[62385]: DEBUG oslo_concurrency.lockutils [req-2a749a3a-4c7a-4b6e-8fdd-18c386882f72 req-ac4b40b9-1475-44d0-84ca-5d03e59407c0 service nova] Lock "f937ef7b-ab39-4d8a-9577-7faabc652aa5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.982582] env[62385]: DEBUG oslo_concurrency.lockutils [req-2a749a3a-4c7a-4b6e-8fdd-18c386882f72 req-ac4b40b9-1475-44d0-84ca-5d03e59407c0 service nova] Lock "f937ef7b-ab39-4d8a-9577-7faabc652aa5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.982808] env[62385]: DEBUG nova.compute.manager [req-2a749a3a-4c7a-4b6e-8fdd-18c386882f72 req-ac4b40b9-1475-44d0-84ca-5d03e59407c0 service nova] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] No waiting events found dispatching network-vif-plugged-2b05b2b6-d4ce-4eb7-9d65-730d4dbad51c {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 893.983047] env[62385]: WARNING nova.compute.manager [req-2a749a3a-4c7a-4b6e-8fdd-18c386882f72 req-ac4b40b9-1475-44d0-84ca-5d03e59407c0 service nova] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Received unexpected event network-vif-plugged-2b05b2b6-d4ce-4eb7-9d65-730d4dbad51c for instance with vm_state building and task_state spawning. [ 893.993412] env[62385]: DEBUG oslo_concurrency.lockutils [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.994409] env[62385]: DEBUG nova.network.neutron [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Successfully updated port: 2b05b2b6-d4ce-4eb7-9d65-730d4dbad51c {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 894.013156] env[62385]: DEBUG nova.scheduler.client.report [None req-6a258a04-6455-49dc-9069-3e3c3f39df66 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 894.038837] env[62385]: DEBUG oslo_vmware.api [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206056, 'name': Rename_Task, 'duration_secs': 0.340791} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.039407] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 894.039407] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-151d1406-e642-4dbd-8151-d745ccc52ebf {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.047200] env[62385]: DEBUG oslo_vmware.api [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 894.047200] env[62385]: value = "task-1206057" [ 894.047200] env[62385]: _type = "Task" [ 894.047200] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.061653] env[62385]: DEBUG oslo_vmware.api [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206057, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.071960] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bff5f4e5-e0ea-4cc7-b233-08032133f4bd tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Lock "2bcd9457-8a73-4e7a-b778-d52c468b3aae" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.251s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.308906] env[62385]: DEBUG nova.compute.manager [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 894.336641] env[62385]: DEBUG nova.compute.manager [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 894.446745] env[62385]: INFO nova.compute.manager [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Rebuilding instance [ 894.496471] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "refresh_cache-f937ef7b-ab39-4d8a-9577-7faabc652aa5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.496621] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquired lock "refresh_cache-f937ef7b-ab39-4d8a-9577-7faabc652aa5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.496768] env[62385]: DEBUG nova.network.neutron [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 894.499696] env[62385]: DEBUG nova.compute.manager [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 894.500631] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14172596-4d9c-4b9f-b840-958b2dcdbea5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.520161] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6a258a04-6455-49dc-9069-3e3c3f39df66 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.916s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.522425] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6040b7c9-146e-4782-b07a-776996b26e1b tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.845s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.522660] env[62385]: DEBUG nova.objects.instance [None req-6040b7c9-146e-4782-b07a-776996b26e1b tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Lazy-loading 'resources' on Instance uuid fbc71ada-d3b5-46f6-90a9-489c118d5126 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 894.540258] env[62385]: INFO nova.scheduler.client.report [None req-6a258a04-6455-49dc-9069-3e3c3f39df66 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Deleted allocations for instance 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c [ 894.558488] env[62385]: DEBUG oslo_vmware.api [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206057, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.721871] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9dbce20d-3a6a-44bc-85e8-6001ac82c9b8 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquiring lock "b7a286a2-9a19-4ef1-8a46-6e2a04e22413" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.722177] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9dbce20d-3a6a-44bc-85e8-6001ac82c9b8 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "b7a286a2-9a19-4ef1-8a46-6e2a04e22413" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.722401] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9dbce20d-3a6a-44bc-85e8-6001ac82c9b8 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquiring lock "b7a286a2-9a19-4ef1-8a46-6e2a04e22413-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.722639] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9dbce20d-3a6a-44bc-85e8-6001ac82c9b8 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "b7a286a2-9a19-4ef1-8a46-6e2a04e22413-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.722846] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9dbce20d-3a6a-44bc-85e8-6001ac82c9b8 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "b7a286a2-9a19-4ef1-8a46-6e2a04e22413-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.725081] env[62385]: INFO nova.compute.manager [None req-9dbce20d-3a6a-44bc-85e8-6001ac82c9b8 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Terminating instance [ 894.726937] env[62385]: DEBUG nova.compute.manager [None req-9dbce20d-3a6a-44bc-85e8-6001ac82c9b8 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 894.727184] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-9dbce20d-3a6a-44bc-85e8-6001ac82c9b8 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 894.728058] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1666878-6d6e-4592-b858-598556c91de7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.736211] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-9dbce20d-3a6a-44bc-85e8-6001ac82c9b8 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 894.736418] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e8550c17-483d-4476-bc48-705835c6bdb5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.743766] env[62385]: DEBUG oslo_vmware.api [None req-9dbce20d-3a6a-44bc-85e8-6001ac82c9b8 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 894.743766] env[62385]: value = "task-1206058" [ 894.743766] env[62385]: _type = "Task" [ 894.743766] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.751974] env[62385]: DEBUG oslo_vmware.api [None req-9dbce20d-3a6a-44bc-85e8-6001ac82c9b8 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206058, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.828571] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.861100] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.013099] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 895.013711] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-231326b0-6f22-41ba-9ef4-7c38d9051540 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.023651] env[62385]: DEBUG oslo_vmware.api [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Waiting for the task: (returnval){ [ 895.023651] env[62385]: value = "task-1206059" [ 895.023651] env[62385]: _type = "Task" [ 895.023651] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.040273] env[62385]: DEBUG oslo_vmware.api [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': task-1206059, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.048064] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6a258a04-6455-49dc-9069-3e3c3f39df66 tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Lock "5e5f43e8-ce45-45c7-a244-bfa724ea5e3c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.112s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.061819] env[62385]: DEBUG oslo_vmware.api [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206057, 'name': PowerOnVM_Task, 'duration_secs': 0.975657} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.063211] env[62385]: DEBUG nova.network.neutron [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 895.065129] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 895.065371] env[62385]: DEBUG nova.compute.manager [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 895.066658] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbc94c89-0292-47c5-941f-c85bd6bf962e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.255469] env[62385]: DEBUG oslo_vmware.api [None req-9dbce20d-3a6a-44bc-85e8-6001ac82c9b8 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206058, 'name': PowerOffVM_Task, 'duration_secs': 0.343572} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.255846] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-9dbce20d-3a6a-44bc-85e8-6001ac82c9b8 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 895.256124] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-9dbce20d-3a6a-44bc-85e8-6001ac82c9b8 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 895.256449] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0df7cce4-090f-42d7-8b7a-b0635f6fc21c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.281788] env[62385]: DEBUG nova.network.neutron [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Updating instance_info_cache with network_info: [{"id": "2b05b2b6-d4ce-4eb7-9d65-730d4dbad51c", "address": "fa:16:3e:ec:d7:16", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b05b2b6-d4", "ovs_interfaceid": "2b05b2b6-d4ce-4eb7-9d65-730d4dbad51c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.336288] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-9dbce20d-3a6a-44bc-85e8-6001ac82c9b8 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 895.336548] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-9dbce20d-3a6a-44bc-85e8-6001ac82c9b8 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 895.336739] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-9dbce20d-3a6a-44bc-85e8-6001ac82c9b8 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Deleting the datastore file [datastore1] b7a286a2-9a19-4ef1-8a46-6e2a04e22413 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 895.339519] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d82650bb-a7e1-4ed9-9990-494d9cc230a3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.347227] env[62385]: DEBUG oslo_vmware.api [None req-9dbce20d-3a6a-44bc-85e8-6001ac82c9b8 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 895.347227] env[62385]: value = "task-1206061" [ 895.347227] env[62385]: _type = "Task" [ 895.347227] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.355896] env[62385]: DEBUG oslo_vmware.api [None req-9dbce20d-3a6a-44bc-85e8-6001ac82c9b8 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206061, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.357325] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c22f974-fb61-47ed-a8b7-4aaff4223999 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.367521] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-212d783a-a70c-419d-97cd-51ea838f9169 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.397438] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14b654bd-7b7c-491f-a269-fef1affeae17 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.404844] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c68e8ca9-51ac-41ba-a715-07496c0e0bcc tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Acquiring lock "b4970e1a-2b29-44e9-b79e-cda3c32dddcd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.405110] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c68e8ca9-51ac-41ba-a715-07496c0e0bcc tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Lock "b4970e1a-2b29-44e9-b79e-cda3c32dddcd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.405389] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c68e8ca9-51ac-41ba-a715-07496c0e0bcc tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Acquiring lock "b4970e1a-2b29-44e9-b79e-cda3c32dddcd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.405504] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c68e8ca9-51ac-41ba-a715-07496c0e0bcc tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Lock "b4970e1a-2b29-44e9-b79e-cda3c32dddcd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.405675] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c68e8ca9-51ac-41ba-a715-07496c0e0bcc tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Lock "b4970e1a-2b29-44e9-b79e-cda3c32dddcd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.408307] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd43250c-82f9-4ed0-8814-e7179e0de178 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.412562] env[62385]: INFO nova.compute.manager [None req-c68e8ca9-51ac-41ba-a715-07496c0e0bcc tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Terminating instance [ 895.414701] env[62385]: DEBUG nova.compute.manager [None req-c68e8ca9-51ac-41ba-a715-07496c0e0bcc tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 895.414903] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c68e8ca9-51ac-41ba-a715-07496c0e0bcc tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 895.415717] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b66bbdc0-a997-4f86-8cdf-3af8a58b7df2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.427651] env[62385]: DEBUG nova.compute.provider_tree [None req-6040b7c9-146e-4782-b07a-776996b26e1b tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 895.434280] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c68e8ca9-51ac-41ba-a715-07496c0e0bcc tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 895.434546] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-23dcee98-0c4d-4904-8800-48506ec43e65 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.442442] env[62385]: DEBUG oslo_vmware.api [None req-c68e8ca9-51ac-41ba-a715-07496c0e0bcc tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for the task: (returnval){ [ 895.442442] env[62385]: value = "task-1206062" [ 895.442442] env[62385]: _type = "Task" [ 895.442442] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.451438] env[62385]: DEBUG oslo_vmware.api [None req-c68e8ca9-51ac-41ba-a715-07496c0e0bcc tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1206062, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.497224] env[62385]: DEBUG oslo_concurrency.lockutils [None req-cfb4ab57-33bf-4b29-b421-079ba2d971bd tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Acquiring lock "d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.497521] env[62385]: DEBUG oslo_concurrency.lockutils [None req-cfb4ab57-33bf-4b29-b421-079ba2d971bd tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.535284] env[62385]: DEBUG oslo_vmware.api [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': task-1206059, 'name': PowerOffVM_Task, 'duration_secs': 0.225748} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.535671] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 895.536059] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 895.536863] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e69d5c0-6c7c-4f9f-9808-5863fc307f62 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.545736] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 895.546049] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-62e18c09-2ff7-4f15-a795-fb861e679451 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.575371] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 895.575546] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Deleting contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 895.575663] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Deleting the datastore file [datastore2] 2bcd9457-8a73-4e7a-b778-d52c468b3aae {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 895.575989] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-578d4a5a-d033-48ca-aa49-aa0660de8808 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.583184] env[62385]: INFO nova.compute.manager [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] bringing vm to original state: 'stopped' [ 895.588630] env[62385]: DEBUG oslo_vmware.api [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Waiting for the task: (returnval){ [ 895.588630] env[62385]: value = "task-1206064" [ 895.588630] env[62385]: _type = "Task" [ 895.588630] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.598234] env[62385]: DEBUG oslo_vmware.api [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': task-1206064, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.784644] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Releasing lock "refresh_cache-f937ef7b-ab39-4d8a-9577-7faabc652aa5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.784998] env[62385]: DEBUG nova.compute.manager [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Instance network_info: |[{"id": "2b05b2b6-d4ce-4eb7-9d65-730d4dbad51c", "address": "fa:16:3e:ec:d7:16", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b05b2b6-d4", "ovs_interfaceid": "2b05b2b6-d4ce-4eb7-9d65-730d4dbad51c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 895.785474] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ec:d7:16', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '685b4083-b748-41fb-a68a-273b1073fa28', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2b05b2b6-d4ce-4eb7-9d65-730d4dbad51c', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 895.793366] env[62385]: DEBUG oslo.service.loopingcall [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 895.793598] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 895.793826] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-94091683-5183-4bec-a070-840dbebcf583 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.815059] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 895.815059] env[62385]: value = "task-1206065" [ 895.815059] env[62385]: _type = "Task" [ 895.815059] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.823809] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206065, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.857807] env[62385]: DEBUG oslo_vmware.api [None req-9dbce20d-3a6a-44bc-85e8-6001ac82c9b8 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206061, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.225702} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.859243] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-9dbce20d-3a6a-44bc-85e8-6001ac82c9b8 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 895.859243] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-9dbce20d-3a6a-44bc-85e8-6001ac82c9b8 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 895.859243] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-9dbce20d-3a6a-44bc-85e8-6001ac82c9b8 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 895.859243] env[62385]: INFO nova.compute.manager [None req-9dbce20d-3a6a-44bc-85e8-6001ac82c9b8 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Took 1.13 seconds to destroy the instance on the hypervisor. [ 895.859243] env[62385]: DEBUG oslo.service.loopingcall [None req-9dbce20d-3a6a-44bc-85e8-6001ac82c9b8 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 895.859243] env[62385]: DEBUG nova.compute.manager [-] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 895.859243] env[62385]: DEBUG nova.network.neutron [-] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 895.930382] env[62385]: DEBUG nova.scheduler.client.report [None req-6040b7c9-146e-4782-b07a-776996b26e1b tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 895.953636] env[62385]: DEBUG oslo_vmware.api [None req-c68e8ca9-51ac-41ba-a715-07496c0e0bcc tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1206062, 'name': PowerOffVM_Task, 'duration_secs': 0.240821} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.953936] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c68e8ca9-51ac-41ba-a715-07496c0e0bcc tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 895.954129] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c68e8ca9-51ac-41ba-a715-07496c0e0bcc tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 895.954390] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1923d946-69d8-4709-bfa2-0958ed7501b1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.000766] env[62385]: INFO nova.compute.manager [None req-cfb4ab57-33bf-4b29-b421-079ba2d971bd tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Detaching volume ddb60eb1-ec31-4439-b8c9-0c53408d8c97 [ 896.010787] env[62385]: DEBUG nova.compute.manager [req-2501ac1a-a75d-400f-87bf-efd790ad3a02 req-290b7a8f-e85c-42c8-a873-4e3cdbd15f8e service nova] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Received event network-changed-2b05b2b6-d4ce-4eb7-9d65-730d4dbad51c {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 896.010997] env[62385]: DEBUG nova.compute.manager [req-2501ac1a-a75d-400f-87bf-efd790ad3a02 req-290b7a8f-e85c-42c8-a873-4e3cdbd15f8e service nova] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Refreshing instance network info cache due to event network-changed-2b05b2b6-d4ce-4eb7-9d65-730d4dbad51c. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 896.011231] env[62385]: DEBUG oslo_concurrency.lockutils [req-2501ac1a-a75d-400f-87bf-efd790ad3a02 req-290b7a8f-e85c-42c8-a873-4e3cdbd15f8e service nova] Acquiring lock "refresh_cache-f937ef7b-ab39-4d8a-9577-7faabc652aa5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.011407] env[62385]: DEBUG oslo_concurrency.lockutils [req-2501ac1a-a75d-400f-87bf-efd790ad3a02 req-290b7a8f-e85c-42c8-a873-4e3cdbd15f8e service nova] Acquired lock "refresh_cache-f937ef7b-ab39-4d8a-9577-7faabc652aa5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.011587] env[62385]: DEBUG nova.network.neutron [req-2501ac1a-a75d-400f-87bf-efd790ad3a02 req-290b7a8f-e85c-42c8-a873-4e3cdbd15f8e service nova] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Refreshing network info cache for port 2b05b2b6-d4ce-4eb7-9d65-730d4dbad51c {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 896.038612] env[62385]: INFO nova.virt.block_device [None req-cfb4ab57-33bf-4b29-b421-079ba2d971bd tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Attempting to driver detach volume ddb60eb1-ec31-4439-b8c9-0c53408d8c97 from mountpoint /dev/sdb [ 896.038850] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfb4ab57-33bf-4b29-b421-079ba2d971bd tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Volume detach. Driver type: vmdk {{(pid=62385) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 896.039010] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfb4ab57-33bf-4b29-b421-079ba2d971bd tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261218', 'volume_id': 'ddb60eb1-ec31-4439-b8c9-0c53408d8c97', 'name': 'volume-ddb60eb1-ec31-4439-b8c9-0c53408d8c97', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd646c9a3-0ed3-4e5b-9fc9-886a1a507f1b', 'attached_at': '', 'detached_at': '', 'volume_id': 'ddb60eb1-ec31-4439-b8c9-0c53408d8c97', 'serial': 'ddb60eb1-ec31-4439-b8c9-0c53408d8c97'} {{(pid=62385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 896.039944] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6c00f6a-d760-440c-b931-a425e672860d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.064807] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-409eb721-229f-4235-b855-e92a94db2cf5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.073243] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c960a3b4-2728-49ed-815f-a596dbf91e56 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.107335] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46c96312-9603-4177-8d18-974cd76fdde2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.110855] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c68e8ca9-51ac-41ba-a715-07496c0e0bcc tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 896.110855] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c68e8ca9-51ac-41ba-a715-07496c0e0bcc tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 896.110855] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c68e8ca9-51ac-41ba-a715-07496c0e0bcc tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Deleting the datastore file [datastore1] b4970e1a-2b29-44e9-b79e-cda3c32dddcd {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 896.110855] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a3136040-65a8-4090-be4a-f9f4e6b0c128 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.120274] env[62385]: DEBUG oslo_vmware.api [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': task-1206064, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.187019} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.134213] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 896.134705] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Deleted contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 896.134812] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 896.138118] env[62385]: DEBUG oslo_vmware.api [None req-c68e8ca9-51ac-41ba-a715-07496c0e0bcc tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for the task: (returnval){ [ 896.138118] env[62385]: value = "task-1206067" [ 896.138118] env[62385]: _type = "Task" [ 896.138118] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.138601] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfb4ab57-33bf-4b29-b421-079ba2d971bd tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] The volume has not been displaced from its original location: [datastore2] volume-ddb60eb1-ec31-4439-b8c9-0c53408d8c97/volume-ddb60eb1-ec31-4439-b8c9-0c53408d8c97.vmdk. No consolidation needed. {{(pid=62385) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 896.144805] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfb4ab57-33bf-4b29-b421-079ba2d971bd tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Reconfiguring VM instance instance-0000003e to detach disk 2001 {{(pid=62385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 896.146327] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-631c0e5b-f792-4a12-ab30-e6c2cdebc47f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.174602] env[62385]: DEBUG oslo_vmware.api [None req-c68e8ca9-51ac-41ba-a715-07496c0e0bcc tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1206067, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.176901] env[62385]: DEBUG oslo_vmware.api [None req-cfb4ab57-33bf-4b29-b421-079ba2d971bd tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for the task: (returnval){ [ 896.176901] env[62385]: value = "task-1206068" [ 896.176901] env[62385]: _type = "Task" [ 896.176901] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.188279] env[62385]: DEBUG oslo_vmware.api [None req-cfb4ab57-33bf-4b29-b421-079ba2d971bd tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206068, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.327076] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206065, 'name': CreateVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.348905] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 896.349288] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 896.435807] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6040b7c9-146e-4782-b07a-776996b26e1b tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.913s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.439512] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.064s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.439512] env[62385]: DEBUG nova.objects.instance [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Lazy-loading 'resources' on Instance uuid f2be792c-a723-4a69-b255-bbe7f0e8692d {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 896.461303] env[62385]: INFO nova.scheduler.client.report [None req-6040b7c9-146e-4782-b07a-776996b26e1b tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Deleted allocations for instance fbc71ada-d3b5-46f6-90a9-489c118d5126 [ 896.602869] env[62385]: DEBUG oslo_concurrency.lockutils [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "d12ffea4-2275-4974-a5c6-b7fa365865b4" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.603134] env[62385]: DEBUG oslo_concurrency.lockutils [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "d12ffea4-2275-4974-a5c6-b7fa365865b4" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.603322] env[62385]: DEBUG nova.compute.manager [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 896.604199] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d53491ff-528f-45b1-b48f-d6a5e1271ecf {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.611723] env[62385]: DEBUG nova.compute.manager [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62385) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 896.614392] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 896.614671] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6731b763-d4f1-46dd-8ec8-36180966f474 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.621043] env[62385]: DEBUG oslo_vmware.api [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 896.621043] env[62385]: value = "task-1206069" [ 896.621043] env[62385]: _type = "Task" [ 896.621043] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.629334] env[62385]: DEBUG oslo_vmware.api [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206069, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.661524] env[62385]: DEBUG oslo_vmware.api [None req-c68e8ca9-51ac-41ba-a715-07496c0e0bcc tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Task: {'id': task-1206067, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.431594} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.661837] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c68e8ca9-51ac-41ba-a715-07496c0e0bcc tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 896.662309] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c68e8ca9-51ac-41ba-a715-07496c0e0bcc tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 896.662309] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c68e8ca9-51ac-41ba-a715-07496c0e0bcc tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 896.662508] env[62385]: INFO nova.compute.manager [None req-c68e8ca9-51ac-41ba-a715-07496c0e0bcc tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Took 1.25 seconds to destroy the instance on the hypervisor. [ 896.662658] env[62385]: DEBUG oslo.service.loopingcall [None req-c68e8ca9-51ac-41ba-a715-07496c0e0bcc tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 896.663091] env[62385]: DEBUG nova.compute.manager [-] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 896.663091] env[62385]: DEBUG nova.network.neutron [-] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 896.686804] env[62385]: DEBUG oslo_vmware.api [None req-cfb4ab57-33bf-4b29-b421-079ba2d971bd tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206068, 'name': ReconfigVM_Task, 'duration_secs': 0.343474} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.687090] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfb4ab57-33bf-4b29-b421-079ba2d971bd tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Reconfigured VM instance instance-0000003e to detach disk 2001 {{(pid=62385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 896.692340] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e72885e5-c2ea-443c-8186-b9010fd488d2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.709016] env[62385]: DEBUG oslo_vmware.api [None req-cfb4ab57-33bf-4b29-b421-079ba2d971bd tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for the task: (returnval){ [ 896.709016] env[62385]: value = "task-1206070" [ 896.709016] env[62385]: _type = "Task" [ 896.709016] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.721298] env[62385]: DEBUG oslo_vmware.api [None req-cfb4ab57-33bf-4b29-b421-079ba2d971bd tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206070, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.728034] env[62385]: DEBUG nova.network.neutron [-] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.749684] env[62385]: DEBUG nova.network.neutron [req-2501ac1a-a75d-400f-87bf-efd790ad3a02 req-290b7a8f-e85c-42c8-a873-4e3cdbd15f8e service nova] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Updated VIF entry in instance network info cache for port 2b05b2b6-d4ce-4eb7-9d65-730d4dbad51c. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 896.750115] env[62385]: DEBUG nova.network.neutron [req-2501ac1a-a75d-400f-87bf-efd790ad3a02 req-290b7a8f-e85c-42c8-a873-4e3cdbd15f8e service nova] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Updating instance_info_cache with network_info: [{"id": "2b05b2b6-d4ce-4eb7-9d65-730d4dbad51c", "address": "fa:16:3e:ec:d7:16", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b05b2b6-d4", "ovs_interfaceid": "2b05b2b6-d4ce-4eb7-9d65-730d4dbad51c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.827384] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206065, 'name': CreateVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.861958] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 896.862278] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Starting heal instance info cache {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 896.970318] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6040b7c9-146e-4782-b07a-776996b26e1b tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Lock "fbc71ada-d3b5-46f6-90a9-489c118d5126" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.369s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.133912] env[62385]: DEBUG oslo_vmware.api [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206069, 'name': PowerOffVM_Task, 'duration_secs': 0.279959} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.137021] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 897.137021] env[62385]: DEBUG nova.compute.manager [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 897.137021] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86742a89-000b-4a6f-939b-c71d3950ba37 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.142106] env[62385]: DEBUG nova.compute.manager [req-d4e94e46-8df8-4a68-a54c-57e74ee9ea1e req-ea37b465-7b0d-43ec-b665-e91d77d6f16c service nova] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Received event network-vif-deleted-854fc77a-662c-4fbe-a30c-5726d1ed6088 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 897.142452] env[62385]: INFO nova.compute.manager [req-d4e94e46-8df8-4a68-a54c-57e74ee9ea1e req-ea37b465-7b0d-43ec-b665-e91d77d6f16c service nova] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Neutron deleted interface 854fc77a-662c-4fbe-a30c-5726d1ed6088; detaching it from the instance and deleting it from the info cache [ 897.142911] env[62385]: DEBUG nova.network.neutron [req-d4e94e46-8df8-4a68-a54c-57e74ee9ea1e req-ea37b465-7b0d-43ec-b665-e91d77d6f16c service nova] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.191911] env[62385]: DEBUG nova.virt.hardware [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 897.191911] env[62385]: DEBUG nova.virt.hardware [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 897.191911] env[62385]: DEBUG nova.virt.hardware [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 897.191911] env[62385]: DEBUG nova.virt.hardware [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 897.192350] env[62385]: DEBUG nova.virt.hardware [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 897.192350] env[62385]: DEBUG nova.virt.hardware [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 897.192350] env[62385]: DEBUG nova.virt.hardware [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 897.193025] env[62385]: DEBUG nova.virt.hardware [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 897.193025] env[62385]: DEBUG nova.virt.hardware [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 897.193025] env[62385]: DEBUG nova.virt.hardware [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 897.193268] env[62385]: DEBUG nova.virt.hardware [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 897.193993] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e09978e4-9377-4023-85aa-ac8a6b0fdc32 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.206152] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6f9db3a-6d69-4247-9c27-4b739e9b5051 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.229587] env[62385]: DEBUG oslo_vmware.api [None req-cfb4ab57-33bf-4b29-b421-079ba2d971bd tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206070, 'name': ReconfigVM_Task, 'duration_secs': 0.231996} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.230104] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Instance VIF info [] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 897.235797] env[62385]: DEBUG oslo.service.loopingcall [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 897.236398] env[62385]: INFO nova.compute.manager [-] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Took 1.38 seconds to deallocate network for instance. [ 897.236972] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfb4ab57-33bf-4b29-b421-079ba2d971bd tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261218', 'volume_id': 'ddb60eb1-ec31-4439-b8c9-0c53408d8c97', 'name': 'volume-ddb60eb1-ec31-4439-b8c9-0c53408d8c97', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd646c9a3-0ed3-4e5b-9fc9-886a1a507f1b', 'attached_at': '', 'detached_at': '', 'volume_id': 'ddb60eb1-ec31-4439-b8c9-0c53408d8c97', 'serial': 'ddb60eb1-ec31-4439-b8c9-0c53408d8c97'} {{(pid=62385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 897.240178] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 897.246407] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-162b832e-3d3a-4be3-914b-9df31b5e8a61 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.271693] env[62385]: DEBUG oslo_concurrency.lockutils [req-2501ac1a-a75d-400f-87bf-efd790ad3a02 req-290b7a8f-e85c-42c8-a873-4e3cdbd15f8e service nova] Releasing lock "refresh_cache-f937ef7b-ab39-4d8a-9577-7faabc652aa5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.280885] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 897.280885] env[62385]: value = "task-1206071" [ 897.280885] env[62385]: _type = "Task" [ 897.280885] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.290029] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206071, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.330813] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206065, 'name': CreateVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.385849] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b372fe43-1f42-4203-bb8b-c2a56567d396 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.394471] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e96a8d35-b1d9-4ee5-bfed-f12d6e1cf76f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.430327] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e443f76-d9d9-42f2-a36f-f8cf10ef8177 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.438661] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38b12a2b-6009-4c59-bacd-6aa7b56c7bc1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.443233] env[62385]: DEBUG nova.network.neutron [-] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.456093] env[62385]: DEBUG nova.compute.provider_tree [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Updating inventory in ProviderTree for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 897.648240] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-41338ce8-afaa-49fd-b448-f127b3484b9d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.657016] env[62385]: DEBUG oslo_concurrency.lockutils [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "d12ffea4-2275-4974-a5c6-b7fa365865b4" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.054s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.660569] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0662d03-18ac-4e89-baf5-f2f5df5dfa8b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.691925] env[62385]: DEBUG nova.compute.manager [req-d4e94e46-8df8-4a68-a54c-57e74ee9ea1e req-ea37b465-7b0d-43ec-b665-e91d77d6f16c service nova] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Detach interface failed, port_id=854fc77a-662c-4fbe-a30c-5726d1ed6088, reason: Instance b4970e1a-2b29-44e9-b79e-cda3c32dddcd could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 897.773469] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9dbce20d-3a6a-44bc-85e8-6001ac82c9b8 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.790905] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206071, 'name': CreateVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.801029] env[62385]: DEBUG nova.objects.instance [None req-cfb4ab57-33bf-4b29-b421-079ba2d971bd tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lazy-loading 'flavor' on Instance uuid d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 897.830680] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206065, 'name': CreateVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.945662] env[62385]: INFO nova.compute.manager [-] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Took 1.28 seconds to deallocate network for instance. [ 897.980611] env[62385]: ERROR nova.scheduler.client.report [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [req-a5ce22b1-2178-465d-807c-f88197cc5e0d] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-a5ce22b1-2178-465d-807c-f88197cc5e0d"}]} [ 897.997847] env[62385]: DEBUG nova.scheduler.client.report [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Refreshing inventories for resource provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 898.013918] env[62385]: DEBUG nova.scheduler.client.report [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Updating ProviderTree inventory for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 898.014181] env[62385]: DEBUG nova.compute.provider_tree [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Updating inventory in ProviderTree for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 898.026863] env[62385]: DEBUG nova.scheduler.client.report [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Refreshing aggregate associations for resource provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a, aggregates: None {{(pid=62385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 898.039338] env[62385]: DEBUG nova.compute.manager [req-01a6ddd3-2b1e-4ba5-9040-6a9267560af4 req-dcb68f53-7d9e-4d86-9f9b-c4358548058e service nova] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Received event network-vif-deleted-2e49e027-be2a-4225-9a34-1e2f9cfb6022 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 898.044764] env[62385]: DEBUG nova.scheduler.client.report [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Refreshing trait associations for resource provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=62385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 898.179244] env[62385]: DEBUG oslo_concurrency.lockutils [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.295784] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206071, 'name': CreateVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.332845] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206065, 'name': CreateVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.388185] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-693bc055-f558-4600-b469-0dcb97a04332 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.397041] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cb1f3e8-e877-45cc-b5ed-2c280d7fa254 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.428932] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquiring lock "refresh_cache-33387505-c576-488b-8c9c-b064fe81a7d7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.429222] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquired lock "refresh_cache-33387505-c576-488b-8c9c-b064fe81a7d7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.429290] env[62385]: DEBUG nova.network.neutron [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Forcefully refreshing network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 898.430944] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2194b5ae-c76b-44ff-9b23-d47671c58ab1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.438956] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a76132c4-89a8-4d8d-81fb-457b162263b9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.453563] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c68e8ca9-51ac-41ba-a715-07496c0e0bcc tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.454070] env[62385]: DEBUG nova.compute.provider_tree [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Updating inventory in ProviderTree for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 898.543290] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0b365ebf-2ef8-461e-bc69-3cbdf7b4242c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "9b9fd167-d843-4b9d-9e4f-dbde5d3d629e" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.543585] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0b365ebf-2ef8-461e-bc69-3cbdf7b4242c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "9b9fd167-d843-4b9d-9e4f-dbde5d3d629e" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.543861] env[62385]: INFO nova.compute.manager [None req-0b365ebf-2ef8-461e-bc69-3cbdf7b4242c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Rebooting instance [ 898.796347] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206071, 'name': CreateVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.807972] env[62385]: DEBUG oslo_concurrency.lockutils [None req-cfb4ab57-33bf-4b29-b421-079ba2d971bd tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.310s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.831657] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206065, 'name': CreateVM_Task, 'duration_secs': 2.932731} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.831825] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 898.832512] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.832682] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.833075] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 898.833342] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8546a88f-3c5e-45b0-97ff-0a6f6813488f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.837918] env[62385]: DEBUG oslo_vmware.api [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 898.837918] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c6edb8-8d70-6b8f-7ead-d9c1d5a45307" [ 898.837918] env[62385]: _type = "Task" [ 898.837918] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.845991] env[62385]: DEBUG oslo_vmware.api [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c6edb8-8d70-6b8f-7ead-d9c1d5a45307, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.856575] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1b4e0f47-b0e4-488a-88a2-73b25e7fe316 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Acquiring lock "d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.856878] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1b4e0f47-b0e4-488a-88a2-73b25e7fe316 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.857110] env[62385]: DEBUG nova.compute.manager [None req-1b4e0f47-b0e4-488a-88a2-73b25e7fe316 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 898.858348] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf9dfdfc-2c25-49fe-a4d4-a061f4c95d43 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.864250] env[62385]: DEBUG nova.compute.manager [None req-1b4e0f47-b0e4-488a-88a2-73b25e7fe316 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62385) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 898.864876] env[62385]: DEBUG nova.objects.instance [None req-1b4e0f47-b0e4-488a-88a2-73b25e7fe316 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lazy-loading 'flavor' on Instance uuid d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 898.990830] env[62385]: DEBUG nova.scheduler.client.report [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Updated inventory for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a with generation 103 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 898.991087] env[62385]: DEBUG nova.compute.provider_tree [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Updating resource provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a generation from 103 to 104 during operation: update_inventory {{(pid=62385) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 898.991250] env[62385]: DEBUG nova.compute.provider_tree [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Updating inventory in ProviderTree for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 899.064539] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0b365ebf-2ef8-461e-bc69-3cbdf7b4242c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "refresh_cache-9b9fd167-d843-4b9d-9e4f-dbde5d3d629e" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 899.064621] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0b365ebf-2ef8-461e-bc69-3cbdf7b4242c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquired lock "refresh_cache-9b9fd167-d843-4b9d-9e4f-dbde5d3d629e" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.064744] env[62385]: DEBUG nova.network.neutron [None req-0b365ebf-2ef8-461e-bc69-3cbdf7b4242c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 899.298035] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206071, 'name': CreateVM_Task, 'duration_secs': 1.583644} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.298035] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 899.298035] env[62385]: DEBUG oslo_concurrency.lockutils [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 899.352581] env[62385]: DEBUG oslo_vmware.api [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c6edb8-8d70-6b8f-7ead-d9c1d5a45307, 'name': SearchDatastore_Task, 'duration_secs': 0.008983} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.352863] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.353111] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 899.353353] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 899.353598] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.355620] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 899.355620] env[62385]: DEBUG oslo_concurrency.lockutils [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.355620] env[62385]: DEBUG oslo_concurrency.lockutils [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 899.355620] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7b7879f1-7626-4339-9271-a96a1fffa103 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.357307] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-213f657c-8502-4ba2-9b9e-e41156aa8c4a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.361398] env[62385]: DEBUG oslo_vmware.api [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Waiting for the task: (returnval){ [ 899.361398] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52647541-242d-07a7-36a7-c0dc275688a2" [ 899.361398] env[62385]: _type = "Task" [ 899.361398] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.364998] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 899.365320] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 899.366417] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1aa4eefb-5bf5-42a9-8d5a-8a8f6458cbe0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.378332] env[62385]: DEBUG oslo_vmware.api [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52647541-242d-07a7-36a7-c0dc275688a2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.379350] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b4e0f47-b0e4-488a-88a2-73b25e7fe316 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 899.379617] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3c9c8555-7380-4842-b80e-59f512283dbc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.383847] env[62385]: DEBUG oslo_vmware.api [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 899.383847] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a41618-39b3-e650-82de-b3f227880ff5" [ 899.383847] env[62385]: _type = "Task" [ 899.383847] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.388838] env[62385]: DEBUG oslo_vmware.api [None req-1b4e0f47-b0e4-488a-88a2-73b25e7fe316 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for the task: (returnval){ [ 899.388838] env[62385]: value = "task-1206072" [ 899.388838] env[62385]: _type = "Task" [ 899.388838] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.396175] env[62385]: DEBUG oslo_vmware.api [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a41618-39b3-e650-82de-b3f227880ff5, 'name': SearchDatastore_Task, 'duration_secs': 0.008861} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.397535] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-157a6744-4fe3-4fd2-8265-7b19938baac2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.403158] env[62385]: DEBUG oslo_vmware.api [None req-1b4e0f47-b0e4-488a-88a2-73b25e7fe316 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206072, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.406331] env[62385]: DEBUG oslo_vmware.api [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 899.406331] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521b3b8b-4ce3-310e-8bf4-2590710124d3" [ 899.406331] env[62385]: _type = "Task" [ 899.406331] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.414358] env[62385]: DEBUG oslo_vmware.api [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521b3b8b-4ce3-310e-8bf4-2590710124d3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.481433] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6287d190-3499-494d-8c3c-9507cccb256b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "d12ffea4-2275-4974-a5c6-b7fa365865b4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.481495] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6287d190-3499-494d-8c3c-9507cccb256b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "d12ffea4-2275-4974-a5c6-b7fa365865b4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.481688] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6287d190-3499-494d-8c3c-9507cccb256b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "d12ffea4-2275-4974-a5c6-b7fa365865b4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.481877] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6287d190-3499-494d-8c3c-9507cccb256b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "d12ffea4-2275-4974-a5c6-b7fa365865b4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.482111] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6287d190-3499-494d-8c3c-9507cccb256b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "d12ffea4-2275-4974-a5c6-b7fa365865b4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.489388] env[62385]: INFO nova.compute.manager [None req-6287d190-3499-494d-8c3c-9507cccb256b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Terminating instance [ 899.489692] env[62385]: DEBUG nova.compute.manager [None req-6287d190-3499-494d-8c3c-9507cccb256b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 899.491185] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-6287d190-3499-494d-8c3c-9507cccb256b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 899.491185] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5f78a44-cbe3-4d0c-8c2b-4c4ef327d9e8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.498609] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.061s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.501396] env[62385]: DEBUG oslo_concurrency.lockutils [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 17.155s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.501447] env[62385]: DEBUG nova.objects.instance [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62385) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 899.509721] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-6287d190-3499-494d-8c3c-9507cccb256b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 899.509721] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-722b0e43-d86f-4eec-9b88-44b52b896214 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.531693] env[62385]: INFO nova.scheduler.client.report [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Deleted allocations for instance f2be792c-a723-4a69-b255-bbe7f0e8692d [ 899.575902] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-6287d190-3499-494d-8c3c-9507cccb256b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 899.576570] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-6287d190-3499-494d-8c3c-9507cccb256b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Deleting contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 899.576570] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-6287d190-3499-494d-8c3c-9507cccb256b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Deleting the datastore file [datastore2] d12ffea4-2275-4974-a5c6-b7fa365865b4 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 899.576723] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3d2027f4-6a65-491b-9215-24054ad130ca {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.584732] env[62385]: DEBUG oslo_vmware.api [None req-6287d190-3499-494d-8c3c-9507cccb256b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 899.584732] env[62385]: value = "task-1206074" [ 899.584732] env[62385]: _type = "Task" [ 899.584732] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.593798] env[62385]: DEBUG oslo_vmware.api [None req-6287d190-3499-494d-8c3c-9507cccb256b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206074, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.674972] env[62385]: DEBUG nova.network.neutron [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Updating instance_info_cache with network_info: [{"id": "72d846e7-a597-4eb7-87ea-0d4309ceb42b", "address": "fa:16:3e:74:68:57", "network": {"id": "b0a3dea4-1925-4920-8a8e-9dc66733d1da", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1951021425-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ac0c8638f064daba042ded4de8c1cc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee20e439-fed9-490e-97dd-f3c886977ae1", "external-id": "nsx-vlan-transportzone-357", "segmentation_id": 357, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72d846e7-a5", "ovs_interfaceid": "72d846e7-a597-4eb7-87ea-0d4309ceb42b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.875857] env[62385]: DEBUG oslo_vmware.api [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52647541-242d-07a7-36a7-c0dc275688a2, 'name': SearchDatastore_Task, 'duration_secs': 0.017851} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.876212] env[62385]: DEBUG oslo_concurrency.lockutils [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.876664] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 899.876939] env[62385]: DEBUG oslo_concurrency.lockutils [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 899.898779] env[62385]: DEBUG oslo_vmware.api [None req-1b4e0f47-b0e4-488a-88a2-73b25e7fe316 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206072, 'name': PowerOffVM_Task, 'duration_secs': 0.181918} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.899065] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b4e0f47-b0e4-488a-88a2-73b25e7fe316 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 899.900183] env[62385]: DEBUG nova.compute.manager [None req-1b4e0f47-b0e4-488a-88a2-73b25e7fe316 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 899.900183] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2f185f7-a942-4c85-9b74-517f86b3451b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.910453] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Acquiring lock "75cf7907-d9e3-4f54-90c9-f8d714e1df40" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.913571] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Lock "75cf7907-d9e3-4f54-90c9-f8d714e1df40" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.928472] env[62385]: DEBUG oslo_vmware.api [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521b3b8b-4ce3-310e-8bf4-2590710124d3, 'name': SearchDatastore_Task, 'duration_secs': 0.009365} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.928873] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.928977] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] f937ef7b-ab39-4d8a-9577-7faabc652aa5/f937ef7b-ab39-4d8a-9577-7faabc652aa5.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 899.929301] env[62385]: DEBUG oslo_concurrency.lockutils [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.929494] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 899.929713] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7d8d7592-35e7-4226-a7c8-2b189f75a18e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.931930] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1ee1bfd9-8e8e-4eb3-814f-8db6d7372bb6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.939507] env[62385]: DEBUG oslo_vmware.api [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 899.939507] env[62385]: value = "task-1206075" [ 899.939507] env[62385]: _type = "Task" [ 899.939507] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.941802] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 899.941802] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 899.944037] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1c400f6-0a86-4a55-93ab-6993ae2f95c4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.955431] env[62385]: DEBUG oslo_vmware.api [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206075, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.956852] env[62385]: DEBUG oslo_vmware.api [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Waiting for the task: (returnval){ [ 899.956852] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]525106d4-476f-6f4c-9fde-1c9376798ebf" [ 899.956852] env[62385]: _type = "Task" [ 899.956852] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.968015] env[62385]: DEBUG oslo_vmware.api [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]525106d4-476f-6f4c-9fde-1c9376798ebf, 'name': SearchDatastore_Task, 'duration_secs': 0.008127} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.968015] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-47ab178d-509d-428c-ba1f-b7f577a6ae53 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.971611] env[62385]: DEBUG nova.network.neutron [None req-0b365ebf-2ef8-461e-bc69-3cbdf7b4242c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Updating instance_info_cache with network_info: [{"id": "404c104c-258f-480e-bfe2-4ace4e83a43c", "address": "fa:16:3e:87:f8:2a", "network": {"id": "06017161-984c-4657-9639-4e6bdb6d4c09", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1072882119-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.208", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d45b5346eab44707b4d81eaaf12b0935", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39cd75b0-9ec7-48ed-b57f-34da0c573a60", "external-id": "nsx-vlan-transportzone-751", "segmentation_id": 751, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap404c104c-25", "ovs_interfaceid": "404c104c-258f-480e-bfe2-4ace4e83a43c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.974757] env[62385]: DEBUG oslo_vmware.api [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Waiting for the task: (returnval){ [ 899.974757] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ab11a4-71e6-09ae-5258-52fcf0e871d9" [ 899.974757] env[62385]: _type = "Task" [ 899.974757] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.984294] env[62385]: DEBUG oslo_vmware.api [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ab11a4-71e6-09ae-5258-52fcf0e871d9, 'name': SearchDatastore_Task, 'duration_secs': 0.007884} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.988018] env[62385]: DEBUG oslo_concurrency.lockutils [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.988018] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 2bcd9457-8a73-4e7a-b778-d52c468b3aae/2bcd9457-8a73-4e7a-b778-d52c468b3aae.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 899.988018] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9d9210da-7fc6-48a9-b454-7965f9ab89ca {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.992948] env[62385]: DEBUG oslo_vmware.api [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Waiting for the task: (returnval){ [ 899.992948] env[62385]: value = "task-1206076" [ 899.992948] env[62385]: _type = "Task" [ 899.992948] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.010681] env[62385]: DEBUG oslo_vmware.api [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': task-1206076, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.040564] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4f2038a7-dde0-425b-ac0f-ddeef4064792 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Lock "f2be792c-a723-4a69-b255-bbe7f0e8692d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.744s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.095160] env[62385]: DEBUG oslo_vmware.api [None req-6287d190-3499-494d-8c3c-9507cccb256b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206074, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133858} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.095422] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-6287d190-3499-494d-8c3c-9507cccb256b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 900.095614] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-6287d190-3499-494d-8c3c-9507cccb256b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Deleted contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 900.095794] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-6287d190-3499-494d-8c3c-9507cccb256b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 900.095970] env[62385]: INFO nova.compute.manager [None req-6287d190-3499-494d-8c3c-9507cccb256b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Took 0.61 seconds to destroy the instance on the hypervisor. [ 900.096246] env[62385]: DEBUG oslo.service.loopingcall [None req-6287d190-3499-494d-8c3c-9507cccb256b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 900.096447] env[62385]: DEBUG nova.compute.manager [-] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 900.096545] env[62385]: DEBUG nova.network.neutron [-] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 900.177741] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Releasing lock "refresh_cache-33387505-c576-488b-8c9c-b064fe81a7d7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.177975] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Updated the network info_cache for instance {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 900.178667] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 900.178893] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 900.179128] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 900.179305] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 900.179457] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 900.179604] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 900.179735] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62385) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 900.179883] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager.update_available_resource {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 900.413028] env[62385]: DEBUG nova.compute.manager [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 900.417926] env[62385]: DEBUG nova.compute.manager [req-baf1c76b-4c28-4106-a267-5e742f758d41 req-7e5ad577-9905-4416-8e04-df2bda81ce82 service nova] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Received event network-vif-deleted-0e719241-7a39-42dd-b37e-31cc75a4da74 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 900.418215] env[62385]: INFO nova.compute.manager [req-baf1c76b-4c28-4106-a267-5e742f758d41 req-7e5ad577-9905-4416-8e04-df2bda81ce82 service nova] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Neutron deleted interface 0e719241-7a39-42dd-b37e-31cc75a4da74; detaching it from the instance and deleting it from the info cache [ 900.418364] env[62385]: DEBUG nova.network.neutron [req-baf1c76b-4c28-4106-a267-5e742f758d41 req-7e5ad577-9905-4416-8e04-df2bda81ce82 service nova] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.427252] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1b4e0f47-b0e4-488a-88a2-73b25e7fe316 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.569s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.456050] env[62385]: DEBUG oslo_vmware.api [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206075, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.43925} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.456050] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] f937ef7b-ab39-4d8a-9577-7faabc652aa5/f937ef7b-ab39-4d8a-9577-7faabc652aa5.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 900.456050] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 900.456441] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c90be7f4-c326-4738-a33a-15d8f04d607f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.466670] env[62385]: DEBUG oslo_vmware.api [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 900.466670] env[62385]: value = "task-1206077" [ 900.466670] env[62385]: _type = "Task" [ 900.466670] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.477972] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0b365ebf-2ef8-461e-bc69-3cbdf7b4242c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Releasing lock "refresh_cache-9b9fd167-d843-4b9d-9e4f-dbde5d3d629e" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.480901] env[62385]: DEBUG oslo_vmware.api [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206077, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.486025] env[62385]: DEBUG nova.compute.manager [None req-0b365ebf-2ef8-461e-bc69-3cbdf7b4242c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 900.486025] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2db4b3ee-fbfa-4a0d-9797-bba64dc8b610 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.509146] env[62385]: DEBUG oslo_vmware.api [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': task-1206076, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.511079] env[62385]: DEBUG oslo_concurrency.lockutils [None req-87d9f9ce-db0d-48a0-9f02-ffe95f784bea tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.512312] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 16.842s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.684592] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.899287] env[62385]: DEBUG nova.network.neutron [-] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.924035] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7d0f3af4-6236-4c63-959b-0ed2729caaf3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.937288] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e64c3a09-f356-42b8-835d-237877bbc789 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.949101] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.970453] env[62385]: DEBUG nova.compute.manager [req-baf1c76b-4c28-4106-a267-5e742f758d41 req-7e5ad577-9905-4416-8e04-df2bda81ce82 service nova] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Detach interface failed, port_id=0e719241-7a39-42dd-b37e-31cc75a4da74, reason: Instance d12ffea4-2275-4974-a5c6-b7fa365865b4 could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 900.979515] env[62385]: DEBUG oslo_vmware.api [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206077, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.192418} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.979862] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 900.980708] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c73f1332-6e65-4c52-b2b7-49fdb6440e24 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.007953] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] f937ef7b-ab39-4d8a-9577-7faabc652aa5/f937ef7b-ab39-4d8a-9577-7faabc652aa5.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 901.014324] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1416a757-2ec9-429a-8f1a-b8ef8020d42d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.040909] env[62385]: INFO nova.compute.claims [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 901.050930] env[62385]: DEBUG oslo_vmware.api [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': task-1206076, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.727064} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.052281] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 2bcd9457-8a73-4e7a-b778-d52c468b3aae/2bcd9457-8a73-4e7a-b778-d52c468b3aae.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 901.052557] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 901.053175] env[62385]: DEBUG oslo_vmware.api [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 901.053175] env[62385]: value = "task-1206078" [ 901.053175] env[62385]: _type = "Task" [ 901.053175] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.053450] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c84c37cf-e058-4baa-84ee-80ffcedee337 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.064915] env[62385]: DEBUG oslo_vmware.api [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206078, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.066197] env[62385]: DEBUG oslo_vmware.api [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Waiting for the task: (returnval){ [ 901.066197] env[62385]: value = "task-1206079" [ 901.066197] env[62385]: _type = "Task" [ 901.066197] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.076340] env[62385]: DEBUG oslo_vmware.api [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': task-1206079, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.190044] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67f292fc-5c39-4e07-ad8d-7dc61472e3a5 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Acquiring lock "e4089e71-eea1-4e16-b90c-97966333f2bc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.190462] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67f292fc-5c39-4e07-ad8d-7dc61472e3a5 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Lock "e4089e71-eea1-4e16-b90c-97966333f2bc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.190692] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67f292fc-5c39-4e07-ad8d-7dc61472e3a5 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Acquiring lock "e4089e71-eea1-4e16-b90c-97966333f2bc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.190874] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67f292fc-5c39-4e07-ad8d-7dc61472e3a5 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Lock "e4089e71-eea1-4e16-b90c-97966333f2bc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.191117] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67f292fc-5c39-4e07-ad8d-7dc61472e3a5 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Lock "e4089e71-eea1-4e16-b90c-97966333f2bc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.193359] env[62385]: INFO nova.compute.manager [None req-67f292fc-5c39-4e07-ad8d-7dc61472e3a5 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Terminating instance [ 901.196542] env[62385]: DEBUG nova.compute.manager [None req-67f292fc-5c39-4e07-ad8d-7dc61472e3a5 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 901.196760] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-67f292fc-5c39-4e07-ad8d-7dc61472e3a5 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 901.197606] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6b82854-a66d-40f2-a291-22c65f91aeb5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.206058] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-67f292fc-5c39-4e07-ad8d-7dc61472e3a5 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 901.206314] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5c15d566-e10c-4821-bfc6-9a42f3bb7aff {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.213801] env[62385]: DEBUG oslo_vmware.api [None req-67f292fc-5c39-4e07-ad8d-7dc61472e3a5 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Waiting for the task: (returnval){ [ 901.213801] env[62385]: value = "task-1206080" [ 901.213801] env[62385]: _type = "Task" [ 901.213801] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.224459] env[62385]: DEBUG oslo_vmware.api [None req-67f292fc-5c39-4e07-ad8d-7dc61472e3a5 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1206080, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.402054] env[62385]: INFO nova.compute.manager [-] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Took 1.31 seconds to deallocate network for instance. [ 901.538816] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1255535-a254-43c7-baa4-c5a65e898569 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.547751] env[62385]: INFO nova.compute.resource_tracker [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Updating resource usage from migration 5c8fd55d-d3a1-49b8-8f53-46b2748a8591 [ 901.550503] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0b365ebf-2ef8-461e-bc69-3cbdf7b4242c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Doing hard reboot of VM {{(pid=62385) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1064}} [ 901.550935] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-b7ab53bc-66fa-4359-a4bb-f3443fb09098 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.557555] env[62385]: DEBUG oslo_vmware.api [None req-0b365ebf-2ef8-461e-bc69-3cbdf7b4242c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 901.557555] env[62385]: value = "task-1206081" [ 901.557555] env[62385]: _type = "Task" [ 901.557555] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.568620] env[62385]: DEBUG oslo_vmware.api [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206078, 'name': ReconfigVM_Task, 'duration_secs': 0.304377} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.576401] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Reconfigured VM instance instance-00000055 to attach disk [datastore1] f937ef7b-ab39-4d8a-9577-7faabc652aa5/f937ef7b-ab39-4d8a-9577-7faabc652aa5.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 901.577061] env[62385]: DEBUG oslo_vmware.api [None req-0b365ebf-2ef8-461e-bc69-3cbdf7b4242c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206081, 'name': ResetVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.577490] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fe26ab1d-db7e-4d56-a113-201b850af5d2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.583513] env[62385]: DEBUG oslo_vmware.api [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': task-1206079, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073791} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.584726] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 901.585082] env[62385]: DEBUG oslo_vmware.api [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 901.585082] env[62385]: value = "task-1206082" [ 901.585082] env[62385]: _type = "Task" [ 901.585082] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.585767] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09a091b1-0f6e-45fc-adeb-a9b83188dfd5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.597987] env[62385]: DEBUG oslo_vmware.api [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206082, 'name': Rename_Task} progress is 10%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.614041] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Reconfiguring VM instance instance-00000054 to attach disk [datastore1] 2bcd9457-8a73-4e7a-b778-d52c468b3aae/2bcd9457-8a73-4e7a-b778-d52c468b3aae.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 901.616878] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d65674cd-c48f-4a3b-b314-4b3a4e775b0e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.635135] env[62385]: DEBUG nova.objects.instance [None req-18e7ad1a-1dc4-4bd4-b330-f10655efc4b4 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lazy-loading 'flavor' on Instance uuid d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 901.638183] env[62385]: DEBUG oslo_vmware.api [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Waiting for the task: (returnval){ [ 901.638183] env[62385]: value = "task-1206083" [ 901.638183] env[62385]: _type = "Task" [ 901.638183] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.650411] env[62385]: DEBUG oslo_vmware.api [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': task-1206083, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.727759] env[62385]: DEBUG oslo_vmware.api [None req-67f292fc-5c39-4e07-ad8d-7dc61472e3a5 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1206080, 'name': PowerOffVM_Task, 'duration_secs': 0.342849} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.728228] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-67f292fc-5c39-4e07-ad8d-7dc61472e3a5 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 901.728228] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-67f292fc-5c39-4e07-ad8d-7dc61472e3a5 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 901.728476] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6a5e0ea4-aff2-4364-9776-0e16875b4a55 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.795584] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-67f292fc-5c39-4e07-ad8d-7dc61472e3a5 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 901.795874] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-67f292fc-5c39-4e07-ad8d-7dc61472e3a5 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Deleting contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 901.796138] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-67f292fc-5c39-4e07-ad8d-7dc61472e3a5 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Deleting the datastore file [datastore2] e4089e71-eea1-4e16-b90c-97966333f2bc {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 901.796436] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-35a903e7-bb42-4dac-9529-614e8963f2aa {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.802767] env[62385]: DEBUG oslo_vmware.api [None req-67f292fc-5c39-4e07-ad8d-7dc61472e3a5 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Waiting for the task: (returnval){ [ 901.802767] env[62385]: value = "task-1206085" [ 901.802767] env[62385]: _type = "Task" [ 901.802767] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.813678] env[62385]: DEBUG oslo_vmware.api [None req-67f292fc-5c39-4e07-ad8d-7dc61472e3a5 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1206085, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.874513] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ab8b5f0-af8d-453c-9855-f7066372924a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.882468] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-128cbbb9-0764-40a8-a9c0-47c530870a00 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.914102] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6287d190-3499-494d-8c3c-9507cccb256b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.915221] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3798b338-ded1-4ce1-b87a-634850d9dc2e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.923471] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-551a4fc3-d8d2-4af9-a1e4-879221ad74ba {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.937361] env[62385]: DEBUG nova.compute.provider_tree [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 902.075437] env[62385]: DEBUG oslo_vmware.api [None req-0b365ebf-2ef8-461e-bc69-3cbdf7b4242c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206081, 'name': ResetVM_Task, 'duration_secs': 0.090007} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.075437] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0b365ebf-2ef8-461e-bc69-3cbdf7b4242c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Did hard reboot of VM {{(pid=62385) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1068}} [ 902.075437] env[62385]: DEBUG nova.compute.manager [None req-0b365ebf-2ef8-461e-bc69-3cbdf7b4242c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 902.075858] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3421152d-d5af-439c-b998-89250dd19c0b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.096139] env[62385]: DEBUG oslo_vmware.api [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206082, 'name': Rename_Task, 'duration_secs': 0.145922} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.096432] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 902.096688] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-24fb3051-5d6b-4c65-b6a5-db785374b1b7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.108032] env[62385]: DEBUG oslo_vmware.api [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 902.108032] env[62385]: value = "task-1206086" [ 902.108032] env[62385]: _type = "Task" [ 902.108032] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.113332] env[62385]: DEBUG oslo_vmware.api [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206086, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.145490] env[62385]: DEBUG oslo_concurrency.lockutils [None req-18e7ad1a-1dc4-4bd4-b330-f10655efc4b4 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Acquiring lock "refresh_cache-d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.145624] env[62385]: DEBUG oslo_concurrency.lockutils [None req-18e7ad1a-1dc4-4bd4-b330-f10655efc4b4 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Acquired lock "refresh_cache-d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.145826] env[62385]: DEBUG nova.network.neutron [None req-18e7ad1a-1dc4-4bd4-b330-f10655efc4b4 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 902.146171] env[62385]: DEBUG nova.objects.instance [None req-18e7ad1a-1dc4-4bd4-b330-f10655efc4b4 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lazy-loading 'info_cache' on Instance uuid d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 902.153395] env[62385]: DEBUG oslo_vmware.api [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': task-1206083, 'name': ReconfigVM_Task, 'duration_secs': 0.335014} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.153598] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Reconfigured VM instance instance-00000054 to attach disk [datastore1] 2bcd9457-8a73-4e7a-b778-d52c468b3aae/2bcd9457-8a73-4e7a-b778-d52c468b3aae.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 902.154219] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e761f210-b033-461b-888a-aa5f48faf0b9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.161363] env[62385]: DEBUG oslo_vmware.api [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Waiting for the task: (returnval){ [ 902.161363] env[62385]: value = "task-1206087" [ 902.161363] env[62385]: _type = "Task" [ 902.161363] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.175461] env[62385]: DEBUG oslo_vmware.api [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': task-1206087, 'name': Rename_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.314125] env[62385]: DEBUG oslo_vmware.api [None req-67f292fc-5c39-4e07-ad8d-7dc61472e3a5 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Task: {'id': task-1206085, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.26553} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.314453] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-67f292fc-5c39-4e07-ad8d-7dc61472e3a5 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 902.314674] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-67f292fc-5c39-4e07-ad8d-7dc61472e3a5 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Deleted contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 902.314887] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-67f292fc-5c39-4e07-ad8d-7dc61472e3a5 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 902.315124] env[62385]: INFO nova.compute.manager [None req-67f292fc-5c39-4e07-ad8d-7dc61472e3a5 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Took 1.12 seconds to destroy the instance on the hypervisor. [ 902.315443] env[62385]: DEBUG oslo.service.loopingcall [None req-67f292fc-5c39-4e07-ad8d-7dc61472e3a5 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 902.316032] env[62385]: DEBUG nova.compute.manager [-] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 902.316032] env[62385]: DEBUG nova.network.neutron [-] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 902.443092] env[62385]: DEBUG nova.scheduler.client.report [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 902.588837] env[62385]: DEBUG nova.compute.manager [req-5e3b677f-ac8c-4d01-8018-28edec69ad7e req-f8de8f11-9fcc-4cfa-b147-858a254f68e0 service nova] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Received event network-vif-deleted-32c6bd5f-7ef8-454d-8360-1bc0fcb5b894 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 902.589130] env[62385]: INFO nova.compute.manager [req-5e3b677f-ac8c-4d01-8018-28edec69ad7e req-f8de8f11-9fcc-4cfa-b147-858a254f68e0 service nova] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Neutron deleted interface 32c6bd5f-7ef8-454d-8360-1bc0fcb5b894; detaching it from the instance and deleting it from the info cache [ 902.589248] env[62385]: DEBUG nova.network.neutron [req-5e3b677f-ac8c-4d01-8018-28edec69ad7e req-f8de8f11-9fcc-4cfa-b147-858a254f68e0 service nova] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.590439] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0b365ebf-2ef8-461e-bc69-3cbdf7b4242c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "9b9fd167-d843-4b9d-9e4f-dbde5d3d629e" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.047s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.615302] env[62385]: DEBUG oslo_vmware.api [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206086, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.649755] env[62385]: DEBUG nova.objects.base [None req-18e7ad1a-1dc4-4bd4-b330-f10655efc4b4 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=62385) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 902.672213] env[62385]: DEBUG oslo_vmware.api [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': task-1206087, 'name': Rename_Task, 'duration_secs': 0.385363} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.673171] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 902.673262] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-38767e13-19d4-4f43-8bf1-3519717ad8e8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.679909] env[62385]: DEBUG oslo_vmware.api [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Waiting for the task: (returnval){ [ 902.679909] env[62385]: value = "task-1206088" [ 902.679909] env[62385]: _type = "Task" [ 902.679909] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.687724] env[62385]: DEBUG oslo_vmware.api [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': task-1206088, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.948197] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.436s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.948431] env[62385]: INFO nova.compute.manager [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Migrating [ 902.948672] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.948826] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquired lock "compute-rpcapi-router" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.950154] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.287s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.952938] env[62385]: INFO nova.compute.claims [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 902.956785] env[62385]: INFO nova.compute.rpcapi [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 902.957297] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Releasing lock "compute-rpcapi-router" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.069330] env[62385]: DEBUG nova.network.neutron [-] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.093372] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d277ad37-730f-4254-9345-0a1fbac54525 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.102647] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c6aee7f-d673-437d-a155-7a2c6d40e319 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.123574] env[62385]: DEBUG oslo_vmware.api [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206086, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.134716] env[62385]: DEBUG nova.compute.manager [req-5e3b677f-ac8c-4d01-8018-28edec69ad7e req-f8de8f11-9fcc-4cfa-b147-858a254f68e0 service nova] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Detach interface failed, port_id=32c6bd5f-7ef8-454d-8360-1bc0fcb5b894, reason: Instance e4089e71-eea1-4e16-b90c-97966333f2bc could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 903.191992] env[62385]: DEBUG oslo_vmware.api [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': task-1206088, 'name': PowerOnVM_Task, 'duration_secs': 0.434068} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.192293] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 903.192517] env[62385]: DEBUG nova.compute.manager [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 903.193278] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d3cf805-4ea8-4075-bc47-9964ce7fe1b8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.475562] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "refresh_cache-a58cb15f-e4df-4b1e-b09c-2f61dce7200f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.475789] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquired lock "refresh_cache-a58cb15f-e4df-4b1e-b09c-2f61dce7200f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.475934] env[62385]: DEBUG nova.network.neutron [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 903.500377] env[62385]: DEBUG nova.network.neutron [None req-18e7ad1a-1dc4-4bd4-b330-f10655efc4b4 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Updating instance_info_cache with network_info: [{"id": "2c31d161-814f-4c23-9882-0aa436bb9e03", "address": "fa:16:3e:d7:43:42", "network": {"id": "aca904e1-fbd6-4def-bbec-bd70c0fc8af9", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-249813742-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.141", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e99f1e4cfd84bf083c8261b2ffcfa2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f7abe5d-b8fe-4983-bd50-e7469f1fe7f3", "external-id": "nsx-vlan-transportzone-263", "segmentation_id": 263, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c31d161-81", "ovs_interfaceid": "2c31d161-814f-4c23-9882-0aa436bb9e03", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.571887] env[62385]: INFO nova.compute.manager [-] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Took 1.26 seconds to deallocate network for instance. [ 903.624162] env[62385]: DEBUG oslo_vmware.api [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206086, 'name': PowerOnVM_Task, 'duration_secs': 1.027993} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.624538] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 903.624618] env[62385]: INFO nova.compute.manager [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Took 10.48 seconds to spawn the instance on the hypervisor. [ 903.624869] env[62385]: DEBUG nova.compute.manager [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 903.625627] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-535b4445-d0f2-4391-8eff-90b88c01d46b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.711064] env[62385]: DEBUG oslo_concurrency.lockutils [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.003213] env[62385]: DEBUG oslo_concurrency.lockutils [None req-18e7ad1a-1dc4-4bd4-b330-f10655efc4b4 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Releasing lock "refresh_cache-d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.079881] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67f292fc-5c39-4e07-ad8d-7dc61472e3a5 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.146377] env[62385]: INFO nova.compute.manager [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Took 33.48 seconds to build instance. [ 904.301160] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f633bfb6-dad3-43eb-9deb-79c920e4d6bb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.310153] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-217fd6fa-8f61-46b0-ad6a-f6e6503492a2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.349510] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ada74c47-0cb2-4e31-bee2-83de1a49589a tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Acquiring lock "2bcd9457-8a73-4e7a-b778-d52c468b3aae" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.349747] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ada74c47-0cb2-4e31-bee2-83de1a49589a tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Lock "2bcd9457-8a73-4e7a-b778-d52c468b3aae" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.349952] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ada74c47-0cb2-4e31-bee2-83de1a49589a tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Acquiring lock "2bcd9457-8a73-4e7a-b778-d52c468b3aae-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.350216] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ada74c47-0cb2-4e31-bee2-83de1a49589a tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Lock "2bcd9457-8a73-4e7a-b778-d52c468b3aae-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.350367] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ada74c47-0cb2-4e31-bee2-83de1a49589a tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Lock "2bcd9457-8a73-4e7a-b778-d52c468b3aae-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.352298] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c948da57-cda4-4012-af34-17c02d54e63b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.355357] env[62385]: INFO nova.compute.manager [None req-ada74c47-0cb2-4e31-bee2-83de1a49589a tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Terminating instance [ 904.357100] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ada74c47-0cb2-4e31-bee2-83de1a49589a tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Acquiring lock "refresh_cache-2bcd9457-8a73-4e7a-b778-d52c468b3aae" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.357267] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ada74c47-0cb2-4e31-bee2-83de1a49589a tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Acquired lock "refresh_cache-2bcd9457-8a73-4e7a-b778-d52c468b3aae" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.357437] env[62385]: DEBUG nova.network.neutron [None req-ada74c47-0cb2-4e31-bee2-83de1a49589a tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 904.363425] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c31cc9f-3428-4560-acfc-33a0825a4d13 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.377592] env[62385]: DEBUG nova.compute.provider_tree [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 904.483540] env[62385]: DEBUG nova.network.neutron [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Updating instance_info_cache with network_info: [{"id": "f5cfcc74-0f00-411e-8ee7-e26ec0a2fe7a", "address": "fa:16:3e:18:18:ff", "network": {"id": "30c4e78b-6a64-47dc-aaaa-a210160c9f23", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1809587656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "105150c170d74e268e82ab894e9bf0d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10ff2092-e8eb-4768-ad4a-65a80560b447", "external-id": "nsx-vlan-transportzone-845", "segmentation_id": 845, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5cfcc74-0f", "ovs_interfaceid": "f5cfcc74-0f00-411e-8ee7-e26ec0a2fe7a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.506734] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-18e7ad1a-1dc4-4bd4-b330-f10655efc4b4 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 904.507034] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5ff59295-c343-4862-8c4a-25d1c1d46d24 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.514806] env[62385]: DEBUG oslo_vmware.api [None req-18e7ad1a-1dc4-4bd4-b330-f10655efc4b4 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for the task: (returnval){ [ 904.514806] env[62385]: value = "task-1206089" [ 904.514806] env[62385]: _type = "Task" [ 904.514806] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.528706] env[62385]: DEBUG oslo_vmware.api [None req-18e7ad1a-1dc4-4bd4-b330-f10655efc4b4 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206089, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.647609] env[62385]: DEBUG oslo_concurrency.lockutils [None req-d5bfdd1a-1f78-4ec5-9ea2-5559b70aae52 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "f937ef7b-ab39-4d8a-9577-7faabc652aa5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.880938] env[62385]: DEBUG nova.scheduler.client.report [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 904.935649] env[62385]: DEBUG nova.network.neutron [None req-ada74c47-0cb2-4e31-bee2-83de1a49589a tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 904.986297] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Releasing lock "refresh_cache-a58cb15f-e4df-4b1e-b09c-2f61dce7200f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.026150] env[62385]: DEBUG oslo_vmware.api [None req-18e7ad1a-1dc4-4bd4-b330-f10655efc4b4 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206089, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.036749] env[62385]: DEBUG nova.network.neutron [None req-ada74c47-0cb2-4e31-bee2-83de1a49589a tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.195440] env[62385]: DEBUG nova.compute.manager [req-d5991343-e365-4639-bd8d-cc2bd88aaecd req-8543f797-5dd9-476f-a227-acb64c89f0a3 service nova] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Received event network-changed-2b05b2b6-d4ce-4eb7-9d65-730d4dbad51c {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 905.195935] env[62385]: DEBUG nova.compute.manager [req-d5991343-e365-4639-bd8d-cc2bd88aaecd req-8543f797-5dd9-476f-a227-acb64c89f0a3 service nova] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Refreshing instance network info cache due to event network-changed-2b05b2b6-d4ce-4eb7-9d65-730d4dbad51c. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 905.196430] env[62385]: DEBUG oslo_concurrency.lockutils [req-d5991343-e365-4639-bd8d-cc2bd88aaecd req-8543f797-5dd9-476f-a227-acb64c89f0a3 service nova] Acquiring lock "refresh_cache-f937ef7b-ab39-4d8a-9577-7faabc652aa5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.196710] env[62385]: DEBUG oslo_concurrency.lockutils [req-d5991343-e365-4639-bd8d-cc2bd88aaecd req-8543f797-5dd9-476f-a227-acb64c89f0a3 service nova] Acquired lock "refresh_cache-f937ef7b-ab39-4d8a-9577-7faabc652aa5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.197051] env[62385]: DEBUG nova.network.neutron [req-d5991343-e365-4639-bd8d-cc2bd88aaecd req-8543f797-5dd9-476f-a227-acb64c89f0a3 service nova] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Refreshing network info cache for port 2b05b2b6-d4ce-4eb7-9d65-730d4dbad51c {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 905.385937] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.436s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.386586] env[62385]: DEBUG nova.compute.manager [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 905.389514] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.133s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.391062] env[62385]: INFO nova.compute.claims [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 905.531668] env[62385]: DEBUG oslo_vmware.api [None req-18e7ad1a-1dc4-4bd4-b330-f10655efc4b4 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206089, 'name': PowerOnVM_Task, 'duration_secs': 0.742999} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.531878] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-18e7ad1a-1dc4-4bd4-b330-f10655efc4b4 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 905.532247] env[62385]: DEBUG nova.compute.manager [None req-18e7ad1a-1dc4-4bd4-b330-f10655efc4b4 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 905.533486] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77ddb3dd-befa-423d-bf7e-f83e0cd26509 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.541357] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ada74c47-0cb2-4e31-bee2-83de1a49589a tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Releasing lock "refresh_cache-2bcd9457-8a73-4e7a-b778-d52c468b3aae" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.541970] env[62385]: DEBUG nova.compute.manager [None req-ada74c47-0cb2-4e31-bee2-83de1a49589a tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 905.542320] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ada74c47-0cb2-4e31-bee2-83de1a49589a tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 905.543719] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e42a5f72-4c98-4d7b-8e22-3e596b56e82b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.554430] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ada74c47-0cb2-4e31-bee2-83de1a49589a tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 905.555091] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-977aa226-8787-4584-99db-2b09a0503793 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.560997] env[62385]: DEBUG oslo_vmware.api [None req-ada74c47-0cb2-4e31-bee2-83de1a49589a tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Waiting for the task: (returnval){ [ 905.560997] env[62385]: value = "task-1206090" [ 905.560997] env[62385]: _type = "Task" [ 905.560997] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.569227] env[62385]: DEBUG oslo_vmware.api [None req-ada74c47-0cb2-4e31-bee2-83de1a49589a tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': task-1206090, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.896471] env[62385]: DEBUG nova.compute.utils [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 905.902823] env[62385]: DEBUG nova.compute.manager [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 905.903227] env[62385]: DEBUG nova.network.neutron [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 905.938620] env[62385]: DEBUG nova.network.neutron [req-d5991343-e365-4639-bd8d-cc2bd88aaecd req-8543f797-5dd9-476f-a227-acb64c89f0a3 service nova] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Updated VIF entry in instance network info cache for port 2b05b2b6-d4ce-4eb7-9d65-730d4dbad51c. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 905.939355] env[62385]: DEBUG nova.network.neutron [req-d5991343-e365-4639-bd8d-cc2bd88aaecd req-8543f797-5dd9-476f-a227-acb64c89f0a3 service nova] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Updating instance_info_cache with network_info: [{"id": "2b05b2b6-d4ce-4eb7-9d65-730d4dbad51c", "address": "fa:16:3e:ec:d7:16", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b05b2b6-d4", "ovs_interfaceid": "2b05b2b6-d4ce-4eb7-9d65-730d4dbad51c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.942861] env[62385]: DEBUG nova.policy [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5e83312d226945c99d05fcc7f74c0978', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bf8e3c942d3445919cfbe988cca84e90', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 906.073376] env[62385]: DEBUG oslo_vmware.api [None req-ada74c47-0cb2-4e31-bee2-83de1a49589a tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': task-1206090, 'name': PowerOffVM_Task, 'duration_secs': 0.169012} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.074398] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ada74c47-0cb2-4e31-bee2-83de1a49589a tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 906.074577] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ada74c47-0cb2-4e31-bee2-83de1a49589a tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 906.074835] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b42121d2-f5f9-4272-8dcf-f2f1a5ed58aa {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.101252] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ada74c47-0cb2-4e31-bee2-83de1a49589a tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 906.102107] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ada74c47-0cb2-4e31-bee2-83de1a49589a tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 906.102107] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ada74c47-0cb2-4e31-bee2-83de1a49589a tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Deleting the datastore file [datastore1] 2bcd9457-8a73-4e7a-b778-d52c468b3aae {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 906.102107] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d99c53be-c182-42be-bdeb-e8411defca23 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.107956] env[62385]: DEBUG oslo_vmware.api [None req-ada74c47-0cb2-4e31-bee2-83de1a49589a tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Waiting for the task: (returnval){ [ 906.107956] env[62385]: value = "task-1206092" [ 906.107956] env[62385]: _type = "Task" [ 906.107956] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.118634] env[62385]: DEBUG oslo_vmware.api [None req-ada74c47-0cb2-4e31-bee2-83de1a49589a tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': task-1206092, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.224296] env[62385]: DEBUG nova.network.neutron [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Successfully created port: 77d05c5a-39f5-45b0-a17c-599e5a2c06fd {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 906.406024] env[62385]: DEBUG nova.compute.manager [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 906.442445] env[62385]: DEBUG oslo_concurrency.lockutils [req-d5991343-e365-4639-bd8d-cc2bd88aaecd req-8543f797-5dd9-476f-a227-acb64c89f0a3 service nova] Releasing lock "refresh_cache-f937ef7b-ab39-4d8a-9577-7faabc652aa5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.512461] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47583eb8-66d4-4bea-b1c9-56cb9cb6cc9b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.533993] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Updating instance 'a58cb15f-e4df-4b1e-b09c-2f61dce7200f' progress to 0 {{(pid=62385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 906.620409] env[62385]: DEBUG oslo_vmware.api [None req-ada74c47-0cb2-4e31-bee2-83de1a49589a tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Task: {'id': task-1206092, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.087696} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.620713] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ada74c47-0cb2-4e31-bee2-83de1a49589a tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 906.620870] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ada74c47-0cb2-4e31-bee2-83de1a49589a tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 906.621064] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ada74c47-0cb2-4e31-bee2-83de1a49589a tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 906.621254] env[62385]: INFO nova.compute.manager [None req-ada74c47-0cb2-4e31-bee2-83de1a49589a tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Took 1.08 seconds to destroy the instance on the hypervisor. [ 906.621518] env[62385]: DEBUG oslo.service.loopingcall [None req-ada74c47-0cb2-4e31-bee2-83de1a49589a tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 906.621715] env[62385]: DEBUG nova.compute.manager [-] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 906.621809] env[62385]: DEBUG nova.network.neutron [-] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 906.642295] env[62385]: DEBUG nova.network.neutron [-] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 906.769054] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dca31d1b-c2a3-4dc0-9b88-5cd471ef11dd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.777789] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b7f07e2-73a8-41b4-95cc-10a89dcf93a0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.809035] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0091f373-af29-408a-8201-f8310b13f0ce {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.816630] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-812586d1-adaa-4228-bfd1-3fe868faa03a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.829646] env[62385]: DEBUG nova.compute.provider_tree [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 907.037920] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 907.038279] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-da71bd67-774b-4a0a-bda2-48730217fb13 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.045891] env[62385]: DEBUG oslo_vmware.api [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 907.045891] env[62385]: value = "task-1206093" [ 907.045891] env[62385]: _type = "Task" [ 907.045891] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.054538] env[62385]: DEBUG oslo_vmware.api [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206093, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.145781] env[62385]: DEBUG nova.network.neutron [-] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.333981] env[62385]: DEBUG nova.scheduler.client.report [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 907.417457] env[62385]: DEBUG nova.compute.manager [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 907.448259] env[62385]: DEBUG nova.virt.hardware [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 907.448597] env[62385]: DEBUG nova.virt.hardware [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 907.448824] env[62385]: DEBUG nova.virt.hardware [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 907.449274] env[62385]: DEBUG nova.virt.hardware [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 907.449347] env[62385]: DEBUG nova.virt.hardware [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 907.449547] env[62385]: DEBUG nova.virt.hardware [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 907.451159] env[62385]: DEBUG nova.virt.hardware [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 907.451159] env[62385]: DEBUG nova.virt.hardware [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 907.451159] env[62385]: DEBUG nova.virt.hardware [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 907.451159] env[62385]: DEBUG nova.virt.hardware [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 907.451159] env[62385]: DEBUG nova.virt.hardware [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 907.451581] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf53bfa3-d893-4c11-ae71-6e225bc7df86 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.459760] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1abac14b-fbdf-4a81-bcec-0266847713cc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.556163] env[62385]: DEBUG oslo_vmware.api [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206093, 'name': PowerOffVM_Task, 'duration_secs': 0.219658} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.556502] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 907.556751] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Updating instance 'a58cb15f-e4df-4b1e-b09c-2f61dce7200f' progress to 17 {{(pid=62385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 907.649382] env[62385]: INFO nova.compute.manager [-] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Took 1.03 seconds to deallocate network for instance. [ 907.845121] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.453s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.845121] env[62385]: DEBUG nova.compute.manager [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 907.846214] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0f12566d-f7e2-4d77-b350-830290273675 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.624s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.846618] env[62385]: DEBUG nova.objects.instance [None req-0f12566d-f7e2-4d77-b350-830290273675 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Lazy-loading 'resources' on Instance uuid 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 907.882310] env[62385]: DEBUG nova.compute.manager [req-adc800ad-8722-4225-9d03-814edc495cfa req-38e16882-4483-4307-b0a2-540aee88b81a service nova] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Received event network-vif-plugged-77d05c5a-39f5-45b0-a17c-599e5a2c06fd {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 907.882310] env[62385]: DEBUG oslo_concurrency.lockutils [req-adc800ad-8722-4225-9d03-814edc495cfa req-38e16882-4483-4307-b0a2-540aee88b81a service nova] Acquiring lock "9b2f8292-9d89-407e-96c5-195ee398cc7b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.883062] env[62385]: DEBUG oslo_concurrency.lockutils [req-adc800ad-8722-4225-9d03-814edc495cfa req-38e16882-4483-4307-b0a2-540aee88b81a service nova] Lock "9b2f8292-9d89-407e-96c5-195ee398cc7b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.883062] env[62385]: DEBUG oslo_concurrency.lockutils [req-adc800ad-8722-4225-9d03-814edc495cfa req-38e16882-4483-4307-b0a2-540aee88b81a service nova] Lock "9b2f8292-9d89-407e-96c5-195ee398cc7b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.883266] env[62385]: DEBUG nova.compute.manager [req-adc800ad-8722-4225-9d03-814edc495cfa req-38e16882-4483-4307-b0a2-540aee88b81a service nova] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] No waiting events found dispatching network-vif-plugged-77d05c5a-39f5-45b0-a17c-599e5a2c06fd {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 907.883426] env[62385]: WARNING nova.compute.manager [req-adc800ad-8722-4225-9d03-814edc495cfa req-38e16882-4483-4307-b0a2-540aee88b81a service nova] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Received unexpected event network-vif-plugged-77d05c5a-39f5-45b0-a17c-599e5a2c06fd for instance with vm_state building and task_state spawning. [ 907.947889] env[62385]: DEBUG nova.network.neutron [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Successfully updated port: 77d05c5a-39f5-45b0-a17c-599e5a2c06fd {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 908.064354] env[62385]: DEBUG nova.virt.hardware [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:16:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 908.064873] env[62385]: DEBUG nova.virt.hardware [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 908.064873] env[62385]: DEBUG nova.virt.hardware [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 908.064973] env[62385]: DEBUG nova.virt.hardware [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 908.065273] env[62385]: DEBUG nova.virt.hardware [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 908.065510] env[62385]: DEBUG nova.virt.hardware [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 908.065740] env[62385]: DEBUG nova.virt.hardware [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 908.065912] env[62385]: DEBUG nova.virt.hardware [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 908.066156] env[62385]: DEBUG nova.virt.hardware [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 908.066375] env[62385]: DEBUG nova.virt.hardware [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 908.066591] env[62385]: DEBUG nova.virt.hardware [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 908.072078] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a2652f2a-048e-4a3c-9871-8e11c7adf3ea {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.091152] env[62385]: DEBUG oslo_vmware.api [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 908.091152] env[62385]: value = "task-1206094" [ 908.091152] env[62385]: _type = "Task" [ 908.091152] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.103208] env[62385]: DEBUG oslo_vmware.api [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206094, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.158151] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ada74c47-0cb2-4e31-bee2-83de1a49589a tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.356101] env[62385]: DEBUG nova.compute.utils [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 908.357776] env[62385]: DEBUG nova.compute.manager [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 908.357952] env[62385]: DEBUG nova.network.neutron [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 908.414232] env[62385]: DEBUG nova.policy [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ed4236fea9984babaf0b48de23a3570d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '212a132650584a4d866dc95004faa391', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 908.450912] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "refresh_cache-9b2f8292-9d89-407e-96c5-195ee398cc7b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.451098] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquired lock "refresh_cache-9b2f8292-9d89-407e-96c5-195ee398cc7b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.451291] env[62385]: DEBUG nova.network.neutron [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 908.604115] env[62385]: DEBUG oslo_vmware.api [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206094, 'name': ReconfigVM_Task, 'duration_secs': 0.261343} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.607075] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Updating instance 'a58cb15f-e4df-4b1e-b09c-2f61dce7200f' progress to 33 {{(pid=62385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 908.688257] env[62385]: DEBUG nova.network.neutron [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Successfully created port: 8c0465dc-0bc0-40bc-bc3b-893bb76bccd7 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 908.694675] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f43ea4c0-49c4-44fc-8a48-155d4004d36f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.703031] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1527c3b-d63c-4074-b22d-6110079f4104 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.736788] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5ed3db9-1b37-40ec-8173-a554a3c2a573 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.744784] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66680512-893c-4d22-bcc8-766513bb2303 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.762028] env[62385]: DEBUG nova.compute.provider_tree [None req-0f12566d-f7e2-4d77-b350-830290273675 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 908.861968] env[62385]: DEBUG nova.compute.manager [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 908.987763] env[62385]: DEBUG nova.network.neutron [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 909.115357] env[62385]: DEBUG nova.virt.hardware [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 909.115649] env[62385]: DEBUG nova.virt.hardware [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 909.115649] env[62385]: DEBUG nova.virt.hardware [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 909.116684] env[62385]: DEBUG nova.virt.hardware [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 909.116684] env[62385]: DEBUG nova.virt.hardware [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 909.116684] env[62385]: DEBUG nova.virt.hardware [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 909.116684] env[62385]: DEBUG nova.virt.hardware [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 909.116684] env[62385]: DEBUG nova.virt.hardware [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 909.116684] env[62385]: DEBUG nova.virt.hardware [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 909.116942] env[62385]: DEBUG nova.virt.hardware [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 909.116942] env[62385]: DEBUG nova.virt.hardware [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 909.122388] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Reconfiguring VM instance instance-00000051 to detach disk 2000 {{(pid=62385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 909.124875] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4140d6e6-53c8-4b02-b187-4b44d0fd3031 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.145667] env[62385]: DEBUG oslo_vmware.api [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 909.145667] env[62385]: value = "task-1206095" [ 909.145667] env[62385]: _type = "Task" [ 909.145667] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.153975] env[62385]: DEBUG oslo_vmware.api [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206095, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.177298] env[62385]: DEBUG nova.network.neutron [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Updating instance_info_cache with network_info: [{"id": "77d05c5a-39f5-45b0-a17c-599e5a2c06fd", "address": "fa:16:3e:c0:1d:33", "network": {"id": "8b2d7a6e-25d4-4a54-ab54-5972612046da", "bridge": "br-int", "label": "tempest-ServersTestJSON-1142134211-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bf8e3c942d3445919cfbe988cca84e90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap77d05c5a-39", "ovs_interfaceid": "77d05c5a-39f5-45b0-a17c-599e5a2c06fd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.261927] env[62385]: DEBUG nova.scheduler.client.report [None req-0f12566d-f7e2-4d77-b350-830290273675 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 909.655872] env[62385]: DEBUG oslo_vmware.api [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206095, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.680013] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Releasing lock "refresh_cache-9b2f8292-9d89-407e-96c5-195ee398cc7b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.680384] env[62385]: DEBUG nova.compute.manager [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Instance network_info: |[{"id": "77d05c5a-39f5-45b0-a17c-599e5a2c06fd", "address": "fa:16:3e:c0:1d:33", "network": {"id": "8b2d7a6e-25d4-4a54-ab54-5972612046da", "bridge": "br-int", "label": "tempest-ServersTestJSON-1142134211-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bf8e3c942d3445919cfbe988cca84e90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap77d05c5a-39", "ovs_interfaceid": "77d05c5a-39f5-45b0-a17c-599e5a2c06fd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 909.680890] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c0:1d:33', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '77d05c5a-39f5-45b0-a17c-599e5a2c06fd', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 909.688465] env[62385]: DEBUG oslo.service.loopingcall [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 909.688768] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 909.688940] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f986ded8-b28f-493a-8d7d-00d4aedb87ea {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.710730] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 909.710730] env[62385]: value = "task-1206096" [ 909.710730] env[62385]: _type = "Task" [ 909.710730] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.718293] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206096, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.767747] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0f12566d-f7e2-4d77-b350-830290273675 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.921s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.771158] env[62385]: DEBUG oslo_concurrency.lockutils [None req-620c0e7e-c879-4917-9593-e5c8b9ecfde9 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.763s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.771158] env[62385]: DEBUG nova.objects.instance [None req-620c0e7e-c879-4917-9593-e5c8b9ecfde9 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Lazy-loading 'resources' on Instance uuid e9bf98d1-089a-4078-b2f8-eb77943731e5 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 909.786797] env[62385]: INFO nova.scheduler.client.report [None req-0f12566d-f7e2-4d77-b350-830290273675 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Deleted allocations for instance 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66 [ 909.872800] env[62385]: DEBUG nova.compute.manager [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 909.899635] env[62385]: DEBUG nova.virt.hardware [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 909.899936] env[62385]: DEBUG nova.virt.hardware [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 909.900124] env[62385]: DEBUG nova.virt.hardware [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 909.900343] env[62385]: DEBUG nova.virt.hardware [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 909.900559] env[62385]: DEBUG nova.virt.hardware [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 909.900779] env[62385]: DEBUG nova.virt.hardware [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 909.901116] env[62385]: DEBUG nova.virt.hardware [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 909.901355] env[62385]: DEBUG nova.virt.hardware [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 909.901601] env[62385]: DEBUG nova.virt.hardware [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 909.901831] env[62385]: DEBUG nova.virt.hardware [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 909.902108] env[62385]: DEBUG nova.virt.hardware [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 909.903361] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e127ff05-aa59-4de9-8ec9-bd4cd3ee3f69 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.911903] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b482931-f5af-40c3-bd6d-a1b87c269663 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.918377] env[62385]: DEBUG nova.compute.manager [req-91c84753-e482-4e76-bb3d-e32c7d58ccde req-2c518e7a-746a-4c0a-a020-8b85283f452f service nova] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Received event network-changed-77d05c5a-39f5-45b0-a17c-599e5a2c06fd {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 909.918589] env[62385]: DEBUG nova.compute.manager [req-91c84753-e482-4e76-bb3d-e32c7d58ccde req-2c518e7a-746a-4c0a-a020-8b85283f452f service nova] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Refreshing instance network info cache due to event network-changed-77d05c5a-39f5-45b0-a17c-599e5a2c06fd. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 909.918826] env[62385]: DEBUG oslo_concurrency.lockutils [req-91c84753-e482-4e76-bb3d-e32c7d58ccde req-2c518e7a-746a-4c0a-a020-8b85283f452f service nova] Acquiring lock "refresh_cache-9b2f8292-9d89-407e-96c5-195ee398cc7b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.918986] env[62385]: DEBUG oslo_concurrency.lockutils [req-91c84753-e482-4e76-bb3d-e32c7d58ccde req-2c518e7a-746a-4c0a-a020-8b85283f452f service nova] Acquired lock "refresh_cache-9b2f8292-9d89-407e-96c5-195ee398cc7b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.919210] env[62385]: DEBUG nova.network.neutron [req-91c84753-e482-4e76-bb3d-e32c7d58ccde req-2c518e7a-746a-4c0a-a020-8b85283f452f service nova] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Refreshing network info cache for port 77d05c5a-39f5-45b0-a17c-599e5a2c06fd {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 910.157617] env[62385]: DEBUG oslo_vmware.api [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206095, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.179091] env[62385]: DEBUG nova.network.neutron [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Successfully updated port: 8c0465dc-0bc0-40bc-bc3b-893bb76bccd7 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 910.221311] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206096, 'name': CreateVM_Task, 'duration_secs': 0.327636} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.221311] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 910.221944] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.222317] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.223021] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 910.223947] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b527ce62-fe35-419b-8898-68116366d1dc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.228382] env[62385]: DEBUG oslo_vmware.api [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 910.228382] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c377b2-7b22-874c-0c31-28ede03a34b2" [ 910.228382] env[62385]: _type = "Task" [ 910.228382] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.236013] env[62385]: DEBUG oslo_vmware.api [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c377b2-7b22-874c-0c31-28ede03a34b2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.298285] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0f12566d-f7e2-4d77-b350-830290273675 tempest-ServersAaction247Test-24441793 tempest-ServersAaction247Test-24441793-project-member] Lock "4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.816s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.543025] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e361a84-f4e3-4f29-819f-45ad238a243a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.550025] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e23003ac-a9e6-4982-8db6-7b1a35218e7a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.586379] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f43ae23f-2005-4fbb-a565-f9d68e78e770 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.595528] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95bfc768-d3d0-4c25-9bcc-91ba3c7e0329 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.607698] env[62385]: DEBUG nova.compute.provider_tree [None req-620c0e7e-c879-4917-9593-e5c8b9ecfde9 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 910.657222] env[62385]: DEBUG oslo_vmware.api [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206095, 'name': ReconfigVM_Task, 'duration_secs': 1.255697} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.657463] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Reconfigured VM instance instance-00000051 to detach disk 2000 {{(pid=62385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 910.658269] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a899985-41ed-4b28-a9fc-43cbdb171a7a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.681298] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] a58cb15f-e4df-4b1e-b09c-2f61dce7200f/a58cb15f-e4df-4b1e-b09c-2f61dce7200f.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 910.684412] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Acquiring lock "refresh_cache-09c017a9-ad18-49b6-b72d-958023c81b24" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.684581] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Acquired lock "refresh_cache-09c017a9-ad18-49b6-b72d-958023c81b24" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.684758] env[62385]: DEBUG nova.network.neutron [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 910.686056] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e3b887d0-fab7-42c9-b21a-b345bea92826 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.711169] env[62385]: DEBUG oslo_vmware.api [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 910.711169] env[62385]: value = "task-1206097" [ 910.711169] env[62385]: _type = "Task" [ 910.711169] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.719684] env[62385]: DEBUG oslo_vmware.api [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206097, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.738027] env[62385]: DEBUG oslo_vmware.api [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c377b2-7b22-874c-0c31-28ede03a34b2, 'name': SearchDatastore_Task, 'duration_secs': 0.009189} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.738909] env[62385]: DEBUG nova.network.neutron [req-91c84753-e482-4e76-bb3d-e32c7d58ccde req-2c518e7a-746a-4c0a-a020-8b85283f452f service nova] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Updated VIF entry in instance network info cache for port 77d05c5a-39f5-45b0-a17c-599e5a2c06fd. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 910.739782] env[62385]: DEBUG nova.network.neutron [req-91c84753-e482-4e76-bb3d-e32c7d58ccde req-2c518e7a-746a-4c0a-a020-8b85283f452f service nova] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Updating instance_info_cache with network_info: [{"id": "77d05c5a-39f5-45b0-a17c-599e5a2c06fd", "address": "fa:16:3e:c0:1d:33", "network": {"id": "8b2d7a6e-25d4-4a54-ab54-5972612046da", "bridge": "br-int", "label": "tempest-ServersTestJSON-1142134211-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bf8e3c942d3445919cfbe988cca84e90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap77d05c5a-39", "ovs_interfaceid": "77d05c5a-39f5-45b0-a17c-599e5a2c06fd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.740488] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.740713] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 910.740948] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.741614] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.741614] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 910.742299] env[62385]: DEBUG nova.network.neutron [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 910.744127] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-df786070-a48f-43a3-a889-b2ec6c522e8c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.752941] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 910.753157] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 910.753883] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f15a4120-289b-4d16-b503-249af4522b91 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.759698] env[62385]: DEBUG oslo_vmware.api [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 910.759698] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]522b0b40-0cbf-8db2-df41-7a787b5bd318" [ 910.759698] env[62385]: _type = "Task" [ 910.759698] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.769710] env[62385]: DEBUG oslo_vmware.api [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]522b0b40-0cbf-8db2-df41-7a787b5bd318, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.975912] env[62385]: DEBUG nova.network.neutron [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Updating instance_info_cache with network_info: [{"id": "8c0465dc-0bc0-40bc-bc3b-893bb76bccd7", "address": "fa:16:3e:4a:dc:5c", "network": {"id": "3325f6ee-8d27-49f8-8959-af54a4bb2230", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1230956501-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "212a132650584a4d866dc95004faa391", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c0465dc-0b", "ovs_interfaceid": "8c0465dc-0bc0-40bc-bc3b-893bb76bccd7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.110755] env[62385]: DEBUG nova.scheduler.client.report [None req-620c0e7e-c879-4917-9593-e5c8b9ecfde9 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 911.221679] env[62385]: DEBUG oslo_vmware.api [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206097, 'name': ReconfigVM_Task, 'duration_secs': 0.352691} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.221987] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Reconfigured VM instance instance-00000051 to attach disk [datastore1] a58cb15f-e4df-4b1e-b09c-2f61dce7200f/a58cb15f-e4df-4b1e-b09c-2f61dce7200f.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 911.222249] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Updating instance 'a58cb15f-e4df-4b1e-b09c-2f61dce7200f' progress to 50 {{(pid=62385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 911.242752] env[62385]: DEBUG oslo_concurrency.lockutils [req-91c84753-e482-4e76-bb3d-e32c7d58ccde req-2c518e7a-746a-4c0a-a020-8b85283f452f service nova] Releasing lock "refresh_cache-9b2f8292-9d89-407e-96c5-195ee398cc7b" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.269918] env[62385]: DEBUG oslo_vmware.api [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]522b0b40-0cbf-8db2-df41-7a787b5bd318, 'name': SearchDatastore_Task, 'duration_secs': 0.008879} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.271372] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-73c6cc9e-88f6-49c6-83eb-6f501f1f6e8f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.276710] env[62385]: DEBUG oslo_vmware.api [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 911.276710] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52497d07-3c6d-c97a-de91-3049273ef47e" [ 911.276710] env[62385]: _type = "Task" [ 911.276710] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.284116] env[62385]: DEBUG oslo_vmware.api [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52497d07-3c6d-c97a-de91-3049273ef47e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.479195] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Releasing lock "refresh_cache-09c017a9-ad18-49b6-b72d-958023c81b24" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.479666] env[62385]: DEBUG nova.compute.manager [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Instance network_info: |[{"id": "8c0465dc-0bc0-40bc-bc3b-893bb76bccd7", "address": "fa:16:3e:4a:dc:5c", "network": {"id": "3325f6ee-8d27-49f8-8959-af54a4bb2230", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1230956501-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "212a132650584a4d866dc95004faa391", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c0465dc-0b", "ovs_interfaceid": "8c0465dc-0bc0-40bc-bc3b-893bb76bccd7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 911.480167] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4a:dc:5c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'db68bd64-5b56-49af-a075-13dcf85cb2e0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8c0465dc-0bc0-40bc-bc3b-893bb76bccd7', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 911.488132] env[62385]: DEBUG oslo.service.loopingcall [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 911.488410] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 911.488653] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-35b4658f-2675-4aae-9ce5-6477368e32f1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.510404] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 911.510404] env[62385]: value = "task-1206098" [ 911.510404] env[62385]: _type = "Task" [ 911.510404] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.517251] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206098, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.615976] env[62385]: DEBUG oslo_concurrency.lockutils [None req-620c0e7e-c879-4917-9593-e5c8b9ecfde9 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.845s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.618411] env[62385]: DEBUG oslo_concurrency.lockutils [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 17.625s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.618610] env[62385]: DEBUG nova.objects.instance [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62385) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 911.645042] env[62385]: INFO nova.scheduler.client.report [None req-620c0e7e-c879-4917-9593-e5c8b9ecfde9 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Deleted allocations for instance e9bf98d1-089a-4078-b2f8-eb77943731e5 [ 911.728806] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d0735eb-be0d-49b9-a74e-8654216b3ea6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.748090] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f7af9db-9955-4924-acfd-1a3361d8e974 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.765905] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Updating instance 'a58cb15f-e4df-4b1e-b09c-2f61dce7200f' progress to 67 {{(pid=62385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 911.788704] env[62385]: DEBUG oslo_vmware.api [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52497d07-3c6d-c97a-de91-3049273ef47e, 'name': SearchDatastore_Task, 'duration_secs': 0.009187} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.788899] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.789194] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 9b2f8292-9d89-407e-96c5-195ee398cc7b/9b2f8292-9d89-407e-96c5-195ee398cc7b.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 911.789461] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c1cc7412-d111-4808-b561-8e6352d81029 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.797683] env[62385]: DEBUG oslo_vmware.api [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 911.797683] env[62385]: value = "task-1206099" [ 911.797683] env[62385]: _type = "Task" [ 911.797683] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.805422] env[62385]: DEBUG oslo_vmware.api [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206099, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.952148] env[62385]: DEBUG nova.compute.manager [req-ecd7d67f-b13f-4bec-b07e-adb91cf8d9cc req-755dd748-a6ec-44e1-8be1-7dc5562f517e service nova] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Received event network-vif-plugged-8c0465dc-0bc0-40bc-bc3b-893bb76bccd7 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 911.952428] env[62385]: DEBUG oslo_concurrency.lockutils [req-ecd7d67f-b13f-4bec-b07e-adb91cf8d9cc req-755dd748-a6ec-44e1-8be1-7dc5562f517e service nova] Acquiring lock "09c017a9-ad18-49b6-b72d-958023c81b24-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.952748] env[62385]: DEBUG oslo_concurrency.lockutils [req-ecd7d67f-b13f-4bec-b07e-adb91cf8d9cc req-755dd748-a6ec-44e1-8be1-7dc5562f517e service nova] Lock "09c017a9-ad18-49b6-b72d-958023c81b24-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.953067] env[62385]: DEBUG oslo_concurrency.lockutils [req-ecd7d67f-b13f-4bec-b07e-adb91cf8d9cc req-755dd748-a6ec-44e1-8be1-7dc5562f517e service nova] Lock "09c017a9-ad18-49b6-b72d-958023c81b24-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.953165] env[62385]: DEBUG nova.compute.manager [req-ecd7d67f-b13f-4bec-b07e-adb91cf8d9cc req-755dd748-a6ec-44e1-8be1-7dc5562f517e service nova] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] No waiting events found dispatching network-vif-plugged-8c0465dc-0bc0-40bc-bc3b-893bb76bccd7 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 911.953363] env[62385]: WARNING nova.compute.manager [req-ecd7d67f-b13f-4bec-b07e-adb91cf8d9cc req-755dd748-a6ec-44e1-8be1-7dc5562f517e service nova] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Received unexpected event network-vif-plugged-8c0465dc-0bc0-40bc-bc3b-893bb76bccd7 for instance with vm_state building and task_state spawning. [ 911.953600] env[62385]: DEBUG nova.compute.manager [req-ecd7d67f-b13f-4bec-b07e-adb91cf8d9cc req-755dd748-a6ec-44e1-8be1-7dc5562f517e service nova] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Received event network-changed-8c0465dc-0bc0-40bc-bc3b-893bb76bccd7 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 911.953710] env[62385]: DEBUG nova.compute.manager [req-ecd7d67f-b13f-4bec-b07e-adb91cf8d9cc req-755dd748-a6ec-44e1-8be1-7dc5562f517e service nova] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Refreshing instance network info cache due to event network-changed-8c0465dc-0bc0-40bc-bc3b-893bb76bccd7. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 911.953864] env[62385]: DEBUG oslo_concurrency.lockutils [req-ecd7d67f-b13f-4bec-b07e-adb91cf8d9cc req-755dd748-a6ec-44e1-8be1-7dc5562f517e service nova] Acquiring lock "refresh_cache-09c017a9-ad18-49b6-b72d-958023c81b24" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.954020] env[62385]: DEBUG oslo_concurrency.lockutils [req-ecd7d67f-b13f-4bec-b07e-adb91cf8d9cc req-755dd748-a6ec-44e1-8be1-7dc5562f517e service nova] Acquired lock "refresh_cache-09c017a9-ad18-49b6-b72d-958023c81b24" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.954236] env[62385]: DEBUG nova.network.neutron [req-ecd7d67f-b13f-4bec-b07e-adb91cf8d9cc req-755dd748-a6ec-44e1-8be1-7dc5562f517e service nova] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Refreshing network info cache for port 8c0465dc-0bc0-40bc-bc3b-893bb76bccd7 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 912.023663] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206098, 'name': CreateVM_Task, 'duration_secs': 0.451188} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.023856] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 912.024788] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.024971] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.026100] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 912.026444] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3f989a53-018b-40d8-945d-1875e3822d17 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.034256] env[62385]: DEBUG oslo_vmware.api [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for the task: (returnval){ [ 912.034256] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ecdc69-cadb-0cf4-0d6d-76f00b3fec56" [ 912.034256] env[62385]: _type = "Task" [ 912.034256] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.047776] env[62385]: DEBUG oslo_vmware.api [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ecdc69-cadb-0cf4-0d6d-76f00b3fec56, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.159193] env[62385]: DEBUG oslo_concurrency.lockutils [None req-620c0e7e-c879-4917-9593-e5c8b9ecfde9 tempest-ServerMetadataTestJSON-33410535 tempest-ServerMetadataTestJSON-33410535-project-member] Lock "e9bf98d1-089a-4078-b2f8-eb77943731e5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.157s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.308442] env[62385]: DEBUG oslo_vmware.api [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206099, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.476965} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.309123] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 9b2f8292-9d89-407e-96c5-195ee398cc7b/9b2f8292-9d89-407e-96c5-195ee398cc7b.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 912.309555] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 912.309975] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-05d71d72-bf9c-4ce9-9341-225dd21ed9c8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.316676] env[62385]: DEBUG oslo_vmware.api [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 912.316676] env[62385]: value = "task-1206100" [ 912.316676] env[62385]: _type = "Task" [ 912.316676] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.326271] env[62385]: DEBUG oslo_vmware.api [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206100, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.327239] env[62385]: DEBUG nova.network.neutron [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Port f5cfcc74-0f00-411e-8ee7-e26ec0a2fe7a binding to destination host cpu-1 is already ACTIVE {{(pid=62385) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 912.545780] env[62385]: DEBUG oslo_vmware.api [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ecdc69-cadb-0cf4-0d6d-76f00b3fec56, 'name': SearchDatastore_Task, 'duration_secs': 0.056662} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.546403] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.546805] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 912.547214] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.547529] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.547857] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 912.548280] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-910fa630-4cc3-4548-94d2-d6e7f531f010 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.558022] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 912.560095] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 912.560095] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6de14802-52c2-492d-bcf4-7ecfe8948f52 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.567121] env[62385]: DEBUG oslo_vmware.api [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for the task: (returnval){ [ 912.567121] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521a6390-f3b0-ec01-5556-bc427e3e7fd4" [ 912.567121] env[62385]: _type = "Task" [ 912.567121] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.575536] env[62385]: DEBUG oslo_vmware.api [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521a6390-f3b0-ec01-5556-bc427e3e7fd4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.631074] env[62385]: DEBUG oslo_concurrency.lockutils [None req-197dccfe-60ee-4511-abce-c5619930ceb0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.631074] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.802s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.632666] env[62385]: INFO nova.compute.claims [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 912.784825] env[62385]: DEBUG nova.network.neutron [req-ecd7d67f-b13f-4bec-b07e-adb91cf8d9cc req-755dd748-a6ec-44e1-8be1-7dc5562f517e service nova] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Updated VIF entry in instance network info cache for port 8c0465dc-0bc0-40bc-bc3b-893bb76bccd7. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 912.784825] env[62385]: DEBUG nova.network.neutron [req-ecd7d67f-b13f-4bec-b07e-adb91cf8d9cc req-755dd748-a6ec-44e1-8be1-7dc5562f517e service nova] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Updating instance_info_cache with network_info: [{"id": "8c0465dc-0bc0-40bc-bc3b-893bb76bccd7", "address": "fa:16:3e:4a:dc:5c", "network": {"id": "3325f6ee-8d27-49f8-8959-af54a4bb2230", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1230956501-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "212a132650584a4d866dc95004faa391", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c0465dc-0b", "ovs_interfaceid": "8c0465dc-0bc0-40bc-bc3b-893bb76bccd7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.835033] env[62385]: DEBUG oslo_vmware.api [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206100, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066689} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.838474] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 912.838474] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ce2aa60-0f16-463b-b3dd-df5408fd5d79 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.866095] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] 9b2f8292-9d89-407e-96c5-195ee398cc7b/9b2f8292-9d89-407e-96c5-195ee398cc7b.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 912.866645] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ae793a9b-b7cb-4445-afeb-0ff468e0c4b7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.890024] env[62385]: DEBUG oslo_vmware.api [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 912.890024] env[62385]: value = "task-1206101" [ 912.890024] env[62385]: _type = "Task" [ 912.890024] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.900352] env[62385]: DEBUG oslo_vmware.api [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206101, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.081353] env[62385]: DEBUG oslo_vmware.api [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521a6390-f3b0-ec01-5556-bc427e3e7fd4, 'name': SearchDatastore_Task, 'duration_secs': 0.022188} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.082886] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-35e83c09-a1bb-4109-8690-c79bf646379c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.088615] env[62385]: DEBUG oslo_vmware.api [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for the task: (returnval){ [ 913.088615] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52928720-41f5-b4df-74bd-219ed0859f23" [ 913.088615] env[62385]: _type = "Task" [ 913.088615] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.100703] env[62385]: DEBUG oslo_vmware.api [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52928720-41f5-b4df-74bd-219ed0859f23, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.287142] env[62385]: DEBUG oslo_concurrency.lockutils [req-ecd7d67f-b13f-4bec-b07e-adb91cf8d9cc req-755dd748-a6ec-44e1-8be1-7dc5562f517e service nova] Releasing lock "refresh_cache-09c017a9-ad18-49b6-b72d-958023c81b24" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.356781] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "a58cb15f-e4df-4b1e-b09c-2f61dce7200f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.357104] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "a58cb15f-e4df-4b1e-b09c-2f61dce7200f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.357311] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "a58cb15f-e4df-4b1e-b09c-2f61dce7200f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.402184] env[62385]: DEBUG oslo_vmware.api [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206101, 'name': ReconfigVM_Task, 'duration_secs': 0.292269} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.402552] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Reconfigured VM instance instance-00000056 to attach disk [datastore1] 9b2f8292-9d89-407e-96c5-195ee398cc7b/9b2f8292-9d89-407e-96c5-195ee398cc7b.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 913.403205] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1f0054a7-e36a-47c0-bde3-a307d54554fa {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.410187] env[62385]: DEBUG oslo_vmware.api [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 913.410187] env[62385]: value = "task-1206102" [ 913.410187] env[62385]: _type = "Task" [ 913.410187] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.424420] env[62385]: DEBUG oslo_vmware.api [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206102, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.601024] env[62385]: DEBUG oslo_vmware.api [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52928720-41f5-b4df-74bd-219ed0859f23, 'name': SearchDatastore_Task, 'duration_secs': 0.009011} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.601532] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.602671] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 09c017a9-ad18-49b6-b72d-958023c81b24/09c017a9-ad18-49b6-b72d-958023c81b24.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 913.603208] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e84a6131-987c-4872-9da8-cb414e909e87 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.611409] env[62385]: DEBUG oslo_vmware.api [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for the task: (returnval){ [ 913.611409] env[62385]: value = "task-1206103" [ 913.611409] env[62385]: _type = "Task" [ 913.611409] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.620233] env[62385]: DEBUG oslo_vmware.api [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1206103, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.923751] env[62385]: DEBUG oslo_vmware.api [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206102, 'name': Rename_Task, 'duration_secs': 0.136787} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.926744] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 913.927343] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-45d13ff1-939a-4df4-aa3d-aa3218601820 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.939035] env[62385]: DEBUG oslo_vmware.api [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 913.939035] env[62385]: value = "task-1206104" [ 913.939035] env[62385]: _type = "Task" [ 913.939035] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.951393] env[62385]: DEBUG oslo_vmware.api [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206104, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.956278] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00e52475-dc1d-4519-ae2f-25ced78f5a43 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.966761] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2dc6bbc-553f-4e85-8ad6-5eb0128575f2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.004833] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50ae1c47-242d-4e70-bf58-26e30a6c7565 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.013421] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55e61793-58e5-4dcf-a333-70f1ca409844 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.030878] env[62385]: DEBUG nova.compute.provider_tree [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 914.127731] env[62385]: DEBUG oslo_vmware.api [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1206103, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.489387} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.128179] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 09c017a9-ad18-49b6-b72d-958023c81b24/09c017a9-ad18-49b6-b72d-958023c81b24.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 914.128548] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 914.128929] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3a2e18ae-77a5-4583-b72d-63a07e49ca5c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.137687] env[62385]: DEBUG oslo_vmware.api [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for the task: (returnval){ [ 914.137687] env[62385]: value = "task-1206105" [ 914.137687] env[62385]: _type = "Task" [ 914.137687] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.151207] env[62385]: DEBUG oslo_vmware.api [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1206105, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.447745] env[62385]: DEBUG oslo_vmware.api [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206104, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.534224] env[62385]: DEBUG nova.scheduler.client.report [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 914.653666] env[62385]: DEBUG oslo_vmware.api [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1206105, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064756} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.654111] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 914.658482] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9d268d4-275d-4de6-8eff-399a64d0c44e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.686615] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Reconfiguring VM instance instance-00000057 to attach disk [datastore1] 09c017a9-ad18-49b6-b72d-958023c81b24/09c017a9-ad18-49b6-b72d-958023c81b24.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 914.687701] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "refresh_cache-a58cb15f-e4df-4b1e-b09c-2f61dce7200f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.687863] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquired lock "refresh_cache-a58cb15f-e4df-4b1e-b09c-2f61dce7200f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.688044] env[62385]: DEBUG nova.network.neutron [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 914.689267] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a9c02e2f-e99b-4bd5-89ab-b812062679d6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.714704] env[62385]: DEBUG oslo_vmware.api [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for the task: (returnval){ [ 914.714704] env[62385]: value = "task-1206106" [ 914.714704] env[62385]: _type = "Task" [ 914.714704] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.725329] env[62385]: DEBUG oslo_vmware.api [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1206106, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.946544] env[62385]: DEBUG oslo_vmware.api [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206104, 'name': PowerOnVM_Task, 'duration_secs': 0.537114} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.946995] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 914.947159] env[62385]: INFO nova.compute.manager [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Took 7.53 seconds to spawn the instance on the hypervisor. [ 914.947259] env[62385]: DEBUG nova.compute.manager [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 914.948123] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-605820a6-3a27-4a18-9e33-9336c309a467 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.040960] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.411s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.041514] env[62385]: DEBUG nova.compute.manager [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 915.043963] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.183s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.045325] env[62385]: INFO nova.compute.claims [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 915.075957] env[62385]: DEBUG nova.network.neutron [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Updating instance_info_cache with network_info: [{"id": "f5cfcc74-0f00-411e-8ee7-e26ec0a2fe7a", "address": "fa:16:3e:18:18:ff", "network": {"id": "30c4e78b-6a64-47dc-aaaa-a210160c9f23", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1809587656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "105150c170d74e268e82ab894e9bf0d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10ff2092-e8eb-4768-ad4a-65a80560b447", "external-id": "nsx-vlan-transportzone-845", "segmentation_id": 845, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5cfcc74-0f", "ovs_interfaceid": "f5cfcc74-0f00-411e-8ee7-e26ec0a2fe7a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.226036] env[62385]: DEBUG oslo_vmware.api [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1206106, 'name': ReconfigVM_Task, 'duration_secs': 0.288413} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.226374] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Reconfigured VM instance instance-00000057 to attach disk [datastore1] 09c017a9-ad18-49b6-b72d-958023c81b24/09c017a9-ad18-49b6-b72d-958023c81b24.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 915.226994] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fc76470d-287e-4268-b9ce-fd60b4bfe7fe {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.233559] env[62385]: DEBUG oslo_vmware.api [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for the task: (returnval){ [ 915.233559] env[62385]: value = "task-1206107" [ 915.233559] env[62385]: _type = "Task" [ 915.233559] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.243168] env[62385]: DEBUG oslo_vmware.api [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1206107, 'name': Rename_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.464412] env[62385]: INFO nova.compute.manager [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Took 30.82 seconds to build instance. [ 915.554957] env[62385]: DEBUG nova.compute.utils [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 915.556399] env[62385]: DEBUG nova.compute.manager [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 915.556621] env[62385]: DEBUG nova.network.neutron [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 915.578454] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Releasing lock "refresh_cache-a58cb15f-e4df-4b1e-b09c-2f61dce7200f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.628864] env[62385]: DEBUG nova.policy [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '12f9521f9c9b4c6b9afc4973be32c5b7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '731b6700ce5b43369b40165545f04950', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 915.747349] env[62385]: DEBUG oslo_vmware.api [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1206107, 'name': Rename_Task, 'duration_secs': 0.141091} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.747658] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 915.750500] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2e5ad048-ff94-4fb9-9402-1d1611b02756 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.758017] env[62385]: DEBUG oslo_vmware.api [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for the task: (returnval){ [ 915.758017] env[62385]: value = "task-1206108" [ 915.758017] env[62385]: _type = "Task" [ 915.758017] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.765022] env[62385]: DEBUG oslo_vmware.api [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1206108, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.866038] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquiring lock "8e1d416c-2624-468c-94d7-6265e9f4178f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.866289] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lock "8e1d416c-2624-468c-94d7-6265e9f4178f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.966983] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bd80f3d2-59bd-4b11-89b1-536075bfd148 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "9b2f8292-9d89-407e-96c5-195ee398cc7b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.334s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.063137] env[62385]: DEBUG nova.compute.manager [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 916.100763] env[62385]: DEBUG nova.network.neutron [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Successfully created port: 43db078b-0071-44a3-ab42-476acd4c56bf {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 916.115022] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-accfe578-a2ec-4897-bcc8-b092e071881c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.143337] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b58007a8-fc86-4bf1-be1b-8949b4c29a9d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.151953] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Updating instance 'a58cb15f-e4df-4b1e-b09c-2f61dce7200f' progress to 83 {{(pid=62385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 916.269285] env[62385]: DEBUG oslo_vmware.api [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1206108, 'name': PowerOnVM_Task, 'duration_secs': 0.458799} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.272171] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 916.272368] env[62385]: INFO nova.compute.manager [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Took 6.40 seconds to spawn the instance on the hypervisor. [ 916.272556] env[62385]: DEBUG nova.compute.manager [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 916.273544] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13529509-7f2d-4789-bf47-d7f4f7a28edf {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.369600] env[62385]: DEBUG nova.compute.manager [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 916.420606] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75369c06-ebff-4e54-8a45-803d6bfd506a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.428691] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2d08534-1c83-4b90-8dde-7c91a99456fa {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.468357] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8699ff0-f13e-4376-a565-76e87722896e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.476111] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-623c005d-f38a-4242-af81-ff84d1ae8abf {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.489703] env[62385]: DEBUG oslo_concurrency.lockutils [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "df559cb8-555c-496b-bce3-0981be3ef65f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.489892] env[62385]: DEBUG oslo_concurrency.lockutils [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "df559cb8-555c-496b-bce3-0981be3ef65f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.499797] env[62385]: DEBUG nova.compute.provider_tree [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Updating inventory in ProviderTree for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 916.659391] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 916.659703] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-02e27f56-ed28-4782-bec8-0609ce566148 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.667677] env[62385]: DEBUG oslo_vmware.api [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 916.667677] env[62385]: value = "task-1206109" [ 916.667677] env[62385]: _type = "Task" [ 916.667677] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.676662] env[62385]: DEBUG oslo_vmware.api [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206109, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.790234] env[62385]: INFO nova.compute.manager [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Took 25.55 seconds to build instance. [ 916.889065] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.001616] env[62385]: DEBUG nova.compute.manager [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 917.025968] env[62385]: ERROR nova.scheduler.client.report [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [req-8e3c0fff-8b36-4d1c-9d53-497f441c3a69] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-8e3c0fff-8b36-4d1c-9d53-497f441c3a69"}]} [ 917.044716] env[62385]: DEBUG nova.scheduler.client.report [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Refreshing inventories for resource provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 917.060785] env[62385]: DEBUG nova.scheduler.client.report [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Updating ProviderTree inventory for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 917.061078] env[62385]: DEBUG nova.compute.provider_tree [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Updating inventory in ProviderTree for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 917.073067] env[62385]: DEBUG nova.compute.manager [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 917.076890] env[62385]: DEBUG nova.scheduler.client.report [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Refreshing aggregate associations for resource provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a, aggregates: None {{(pid=62385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 917.102038] env[62385]: DEBUG nova.scheduler.client.report [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Refreshing trait associations for resource provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=62385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 917.111191] env[62385]: DEBUG nova.virt.hardware [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='f8bfafa78b0917fca696d16f4e7c328f',container_format='bare',created_at=2024-10-09T02:23:46Z,direct_url=,disk_format='vmdk',id=aa5b9007-7bfe-4917-a311-121e4cac3174,min_disk=1,min_ram=0,name='tempest-test-snap-2004966963',owner='731b6700ce5b43369b40165545f04950',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-09T02:24:02Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 917.111488] env[62385]: DEBUG nova.virt.hardware [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 917.111766] env[62385]: DEBUG nova.virt.hardware [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 917.111885] env[62385]: DEBUG nova.virt.hardware [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 917.112066] env[62385]: DEBUG nova.virt.hardware [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 917.112239] env[62385]: DEBUG nova.virt.hardware [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 917.112475] env[62385]: DEBUG nova.virt.hardware [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 917.112655] env[62385]: DEBUG nova.virt.hardware [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 917.112852] env[62385]: DEBUG nova.virt.hardware [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 917.113046] env[62385]: DEBUG nova.virt.hardware [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 917.113255] env[62385]: DEBUG nova.virt.hardware [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 917.114497] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcc29cc7-95a3-43d0-9820-6c73b8849ec7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.127927] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90f1e54f-ad31-47bf-8c84-d3469c929671 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.180140] env[62385]: DEBUG oslo_vmware.api [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206109, 'name': PowerOnVM_Task, 'duration_secs': 0.460534} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.180433] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 917.180626] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fef00fc6-979a-41cc-b911-079bdd3ee328 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Updating instance 'a58cb15f-e4df-4b1e-b09c-2f61dce7200f' progress to 100 {{(pid=62385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 917.291428] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec0cbfcd-f29e-4156-b2d9-24a582f4a099 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Lock "09c017a9-ad18-49b6-b72d-958023c81b24" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.061s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.434244] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2c77611-0946-4cca-be8f-1ff7e12edce3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.442384] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a08434f-0bae-4ca1-afbe-8ce1b7a28cd3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.473544] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60befe6f-42e4-4390-bde5-52b362ea6f0d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.480938] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ece98993-a5d4-4be3-bc74-384272c9c685 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.495258] env[62385]: DEBUG nova.compute.provider_tree [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Updating inventory in ProviderTree for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 917.524038] env[62385]: DEBUG oslo_concurrency.lockutils [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.705285] env[62385]: DEBUG nova.compute.manager [req-712700d5-a1dd-4db3-aae3-0562d6c31e1a req-8060f9e8-0962-4fa1-b403-130299046b06 service nova] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Received event network-vif-plugged-43db078b-0071-44a3-ab42-476acd4c56bf {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 917.705343] env[62385]: DEBUG oslo_concurrency.lockutils [req-712700d5-a1dd-4db3-aae3-0562d6c31e1a req-8060f9e8-0962-4fa1-b403-130299046b06 service nova] Acquiring lock "caab700b-bd3c-497e-b40a-eb9d20dfc1e2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.705525] env[62385]: DEBUG oslo_concurrency.lockutils [req-712700d5-a1dd-4db3-aae3-0562d6c31e1a req-8060f9e8-0962-4fa1-b403-130299046b06 service nova] Lock "caab700b-bd3c-497e-b40a-eb9d20dfc1e2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.705694] env[62385]: DEBUG oslo_concurrency.lockutils [req-712700d5-a1dd-4db3-aae3-0562d6c31e1a req-8060f9e8-0962-4fa1-b403-130299046b06 service nova] Lock "caab700b-bd3c-497e-b40a-eb9d20dfc1e2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.705863] env[62385]: DEBUG nova.compute.manager [req-712700d5-a1dd-4db3-aae3-0562d6c31e1a req-8060f9e8-0962-4fa1-b403-130299046b06 service nova] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] No waiting events found dispatching network-vif-plugged-43db078b-0071-44a3-ab42-476acd4c56bf {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 917.706068] env[62385]: WARNING nova.compute.manager [req-712700d5-a1dd-4db3-aae3-0562d6c31e1a req-8060f9e8-0962-4fa1-b403-130299046b06 service nova] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Received unexpected event network-vif-plugged-43db078b-0071-44a3-ab42-476acd4c56bf for instance with vm_state building and task_state spawning. [ 917.799117] env[62385]: DEBUG nova.network.neutron [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Successfully updated port: 43db078b-0071-44a3-ab42-476acd4c56bf {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 917.820119] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c0464272-cede-4d75-8256-7a60bb9810a1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Acquiring lock "09c017a9-ad18-49b6-b72d-958023c81b24" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.820403] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c0464272-cede-4d75-8256-7a60bb9810a1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Lock "09c017a9-ad18-49b6-b72d-958023c81b24" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.820618] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c0464272-cede-4d75-8256-7a60bb9810a1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Acquiring lock "09c017a9-ad18-49b6-b72d-958023c81b24-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.820806] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c0464272-cede-4d75-8256-7a60bb9810a1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Lock "09c017a9-ad18-49b6-b72d-958023c81b24-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.820993] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c0464272-cede-4d75-8256-7a60bb9810a1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Lock "09c017a9-ad18-49b6-b72d-958023c81b24-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.823638] env[62385]: INFO nova.compute.manager [None req-c0464272-cede-4d75-8256-7a60bb9810a1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Terminating instance [ 917.825406] env[62385]: DEBUG nova.compute.manager [None req-c0464272-cede-4d75-8256-7a60bb9810a1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 917.825603] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c0464272-cede-4d75-8256-7a60bb9810a1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 917.826455] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0bf4a3f-d07c-4381-812e-44ea21ddd53d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.836428] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0464272-cede-4d75-8256-7a60bb9810a1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 917.837139] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ee847d47-47ed-4afa-9cc8-542aadd2df97 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.844437] env[62385]: DEBUG oslo_vmware.api [None req-c0464272-cede-4d75-8256-7a60bb9810a1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for the task: (returnval){ [ 917.844437] env[62385]: value = "task-1206110" [ 917.844437] env[62385]: _type = "Task" [ 917.844437] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.852513] env[62385]: DEBUG oslo_vmware.api [None req-c0464272-cede-4d75-8256-7a60bb9810a1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1206110, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.026497] env[62385]: DEBUG nova.scheduler.client.report [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Updated inventory for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a with generation 107 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 918.026796] env[62385]: DEBUG nova.compute.provider_tree [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Updating resource provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a generation from 107 to 108 during operation: update_inventory {{(pid=62385) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 918.026982] env[62385]: DEBUG nova.compute.provider_tree [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Updating inventory in ProviderTree for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 918.304292] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquiring lock "refresh_cache-caab700b-bd3c-497e-b40a-eb9d20dfc1e2" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.304559] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquired lock "refresh_cache-caab700b-bd3c-497e-b40a-eb9d20dfc1e2" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.304739] env[62385]: DEBUG nova.network.neutron [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 918.355314] env[62385]: DEBUG oslo_vmware.api [None req-c0464272-cede-4d75-8256-7a60bb9810a1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1206110, 'name': PowerOffVM_Task, 'duration_secs': 0.392652} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.355578] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0464272-cede-4d75-8256-7a60bb9810a1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 918.355757] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c0464272-cede-4d75-8256-7a60bb9810a1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 918.356119] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-28a725f3-6e27-47c3-80f7-7d8baa1f3806 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.427649] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c0464272-cede-4d75-8256-7a60bb9810a1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 918.427883] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c0464272-cede-4d75-8256-7a60bb9810a1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 918.428166] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0464272-cede-4d75-8256-7a60bb9810a1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Deleting the datastore file [datastore1] 09c017a9-ad18-49b6-b72d-958023c81b24 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 918.428432] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2e4ab316-98c5-48d3-9c19-c87309c9040f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.436483] env[62385]: DEBUG oslo_vmware.api [None req-c0464272-cede-4d75-8256-7a60bb9810a1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for the task: (returnval){ [ 918.436483] env[62385]: value = "task-1206112" [ 918.436483] env[62385]: _type = "Task" [ 918.436483] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.444888] env[62385]: DEBUG oslo_vmware.api [None req-c0464272-cede-4d75-8256-7a60bb9810a1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1206112, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.532142] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.488s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.532694] env[62385]: DEBUG nova.compute.manager [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 918.535394] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9dbce20d-3a6a-44bc-85e8-6001ac82c9b8 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.762s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.535635] env[62385]: DEBUG nova.objects.instance [None req-9dbce20d-3a6a-44bc-85e8-6001ac82c9b8 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lazy-loading 'resources' on Instance uuid b7a286a2-9a19-4ef1-8a46-6e2a04e22413 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 918.837663] env[62385]: DEBUG nova.network.neutron [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 918.927296] env[62385]: DEBUG oslo_concurrency.lockutils [None req-53444455-1881-4ba7-b70c-4c868d7f0a6e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "a58cb15f-e4df-4b1e-b09c-2f61dce7200f" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.927637] env[62385]: DEBUG oslo_concurrency.lockutils [None req-53444455-1881-4ba7-b70c-4c868d7f0a6e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "a58cb15f-e4df-4b1e-b09c-2f61dce7200f" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.927867] env[62385]: DEBUG nova.compute.manager [None req-53444455-1881-4ba7-b70c-4c868d7f0a6e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Going to confirm migration 1 {{(pid=62385) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 918.945946] env[62385]: DEBUG oslo_vmware.api [None req-c0464272-cede-4d75-8256-7a60bb9810a1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Task: {'id': task-1206112, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.214582} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.946220] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0464272-cede-4d75-8256-7a60bb9810a1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 918.946404] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c0464272-cede-4d75-8256-7a60bb9810a1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 918.946586] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c0464272-cede-4d75-8256-7a60bb9810a1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 918.946765] env[62385]: INFO nova.compute.manager [None req-c0464272-cede-4d75-8256-7a60bb9810a1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Took 1.12 seconds to destroy the instance on the hypervisor. [ 918.947011] env[62385]: DEBUG oslo.service.loopingcall [None req-c0464272-cede-4d75-8256-7a60bb9810a1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 918.947208] env[62385]: DEBUG nova.compute.manager [-] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 918.947308] env[62385]: DEBUG nova.network.neutron [-] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 919.025954] env[62385]: DEBUG nova.network.neutron [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Updating instance_info_cache with network_info: [{"id": "43db078b-0071-44a3-ab42-476acd4c56bf", "address": "fa:16:3e:6f:c6:34", "network": {"id": "2c1e2031-ab59-47f5-a758-3a314aca6c1e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1100903482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "731b6700ce5b43369b40165545f04950", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1316f5aa-529f-4bac-8dd7-6076a9d43312", "external-id": "nsx-vlan-transportzone-399", "segmentation_id": 399, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43db078b-00", "ovs_interfaceid": "43db078b-0071-44a3-ab42-476acd4c56bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.042162] env[62385]: DEBUG nova.compute.utils [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 919.045735] env[62385]: DEBUG nova.compute.manager [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 919.045735] env[62385]: DEBUG nova.network.neutron [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 919.083960] env[62385]: DEBUG nova.policy [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e1200cdf75b04233b4c3f4626d4e27cf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aea48968d4464ac0ab01c998c059107d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 919.232966] env[62385]: DEBUG nova.compute.manager [req-b1c4e5bf-e9de-4342-bc0b-562d3ae52759 req-87d87864-a190-4a92-9548-c335b4c37e0a service nova] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Received event network-vif-deleted-8c0465dc-0bc0-40bc-bc3b-893bb76bccd7 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 919.233199] env[62385]: INFO nova.compute.manager [req-b1c4e5bf-e9de-4342-bc0b-562d3ae52759 req-87d87864-a190-4a92-9548-c335b4c37e0a service nova] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Neutron deleted interface 8c0465dc-0bc0-40bc-bc3b-893bb76bccd7; detaching it from the instance and deleting it from the info cache [ 919.233377] env[62385]: DEBUG nova.network.neutron [req-b1c4e5bf-e9de-4342-bc0b-562d3ae52759 req-87d87864-a190-4a92-9548-c335b4c37e0a service nova] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.371500] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b9ea2f7-f712-4a0f-868d-6549f3a196ec {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.380021] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd6f2253-58e5-49d8-b307-f0ae5079ccdf {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.410504] env[62385]: DEBUG nova.network.neutron [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Successfully created port: f09e5aed-7121-4aae-b8e1-e6d84339486f {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 919.412851] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac1b28ad-adcc-4852-9603-beb3e0840ee5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.420167] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-612217ce-663f-4f1c-8910-b4c75f1a97f7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.436131] env[62385]: DEBUG nova.compute.provider_tree [None req-9dbce20d-3a6a-44bc-85e8-6001ac82c9b8 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 919.516839] env[62385]: DEBUG oslo_concurrency.lockutils [None req-53444455-1881-4ba7-b70c-4c868d7f0a6e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "refresh_cache-a58cb15f-e4df-4b1e-b09c-2f61dce7200f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.516839] env[62385]: DEBUG oslo_concurrency.lockutils [None req-53444455-1881-4ba7-b70c-4c868d7f0a6e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquired lock "refresh_cache-a58cb15f-e4df-4b1e-b09c-2f61dce7200f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.517172] env[62385]: DEBUG nova.network.neutron [None req-53444455-1881-4ba7-b70c-4c868d7f0a6e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 919.517274] env[62385]: DEBUG nova.objects.instance [None req-53444455-1881-4ba7-b70c-4c868d7f0a6e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lazy-loading 'info_cache' on Instance uuid a58cb15f-e4df-4b1e-b09c-2f61dce7200f {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 919.528615] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Releasing lock "refresh_cache-caab700b-bd3c-497e-b40a-eb9d20dfc1e2" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.528915] env[62385]: DEBUG nova.compute.manager [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Instance network_info: |[{"id": "43db078b-0071-44a3-ab42-476acd4c56bf", "address": "fa:16:3e:6f:c6:34", "network": {"id": "2c1e2031-ab59-47f5-a758-3a314aca6c1e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1100903482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "731b6700ce5b43369b40165545f04950", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1316f5aa-529f-4bac-8dd7-6076a9d43312", "external-id": "nsx-vlan-transportzone-399", "segmentation_id": 399, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43db078b-00", "ovs_interfaceid": "43db078b-0071-44a3-ab42-476acd4c56bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 919.529409] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6f:c6:34', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1316f5aa-529f-4bac-8dd7-6076a9d43312', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '43db078b-0071-44a3-ab42-476acd4c56bf', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 919.541074] env[62385]: DEBUG oslo.service.loopingcall [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 919.541996] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 919.542859] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-feea8cf3-d79d-481f-8d4c-11de98688689 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.557812] env[62385]: DEBUG nova.compute.manager [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 919.567345] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 919.567345] env[62385]: value = "task-1206113" [ 919.567345] env[62385]: _type = "Task" [ 919.567345] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.577846] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206113, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.704982] env[62385]: DEBUG nova.network.neutron [-] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.737474] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e3d6567f-925f-4513-b324-40e192a8839c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.747293] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-070cb9ec-3c47-4d57-ae87-8280358391d1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.777932] env[62385]: DEBUG nova.compute.manager [req-b1c4e5bf-e9de-4342-bc0b-562d3ae52759 req-87d87864-a190-4a92-9548-c335b4c37e0a service nova] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Detach interface failed, port_id=8c0465dc-0bc0-40bc-bc3b-893bb76bccd7, reason: Instance 09c017a9-ad18-49b6-b72d-958023c81b24 could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 919.813155] env[62385]: DEBUG nova.compute.manager [req-30a9c68b-faea-480e-b200-143c21a5058f req-47a8b430-815b-402b-853c-5043185c2720 service nova] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Received event network-changed-43db078b-0071-44a3-ab42-476acd4c56bf {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 919.813155] env[62385]: DEBUG nova.compute.manager [req-30a9c68b-faea-480e-b200-143c21a5058f req-47a8b430-815b-402b-853c-5043185c2720 service nova] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Refreshing instance network info cache due to event network-changed-43db078b-0071-44a3-ab42-476acd4c56bf. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 919.813155] env[62385]: DEBUG oslo_concurrency.lockutils [req-30a9c68b-faea-480e-b200-143c21a5058f req-47a8b430-815b-402b-853c-5043185c2720 service nova] Acquiring lock "refresh_cache-caab700b-bd3c-497e-b40a-eb9d20dfc1e2" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.813155] env[62385]: DEBUG oslo_concurrency.lockutils [req-30a9c68b-faea-480e-b200-143c21a5058f req-47a8b430-815b-402b-853c-5043185c2720 service nova] Acquired lock "refresh_cache-caab700b-bd3c-497e-b40a-eb9d20dfc1e2" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.813756] env[62385]: DEBUG nova.network.neutron [req-30a9c68b-faea-480e-b200-143c21a5058f req-47a8b430-815b-402b-853c-5043185c2720 service nova] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Refreshing network info cache for port 43db078b-0071-44a3-ab42-476acd4c56bf {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 919.940815] env[62385]: DEBUG nova.scheduler.client.report [None req-9dbce20d-3a6a-44bc-85e8-6001ac82c9b8 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 920.078732] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206113, 'name': CreateVM_Task, 'duration_secs': 0.324417} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.078732] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 920.079609] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/aa5b9007-7bfe-4917-a311-121e4cac3174" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.080396] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquired lock "[datastore1] devstack-image-cache_base/aa5b9007-7bfe-4917-a311-121e4cac3174" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.080396] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/aa5b9007-7bfe-4917-a311-121e4cac3174" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 920.080537] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b914089-3b36-4773-97cc-47e15842fe83 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.085069] env[62385]: DEBUG oslo_vmware.api [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 920.085069] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c6f981-b6e2-8f64-b7c7-e5f36e78f462" [ 920.085069] env[62385]: _type = "Task" [ 920.085069] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.095367] env[62385]: DEBUG oslo_vmware.api [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c6f981-b6e2-8f64-b7c7-e5f36e78f462, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.208742] env[62385]: INFO nova.compute.manager [-] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Took 1.26 seconds to deallocate network for instance. [ 920.446325] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9dbce20d-3a6a-44bc-85e8-6001ac82c9b8 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.911s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.451110] env[62385]: DEBUG oslo_concurrency.lockutils [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 22.272s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.451323] env[62385]: DEBUG nova.objects.instance [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62385) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 920.478561] env[62385]: INFO nova.scheduler.client.report [None req-9dbce20d-3a6a-44bc-85e8-6001ac82c9b8 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Deleted allocations for instance b7a286a2-9a19-4ef1-8a46-6e2a04e22413 [ 920.526060] env[62385]: DEBUG nova.network.neutron [req-30a9c68b-faea-480e-b200-143c21a5058f req-47a8b430-815b-402b-853c-5043185c2720 service nova] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Updated VIF entry in instance network info cache for port 43db078b-0071-44a3-ab42-476acd4c56bf. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 920.526410] env[62385]: DEBUG nova.network.neutron [req-30a9c68b-faea-480e-b200-143c21a5058f req-47a8b430-815b-402b-853c-5043185c2720 service nova] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Updating instance_info_cache with network_info: [{"id": "43db078b-0071-44a3-ab42-476acd4c56bf", "address": "fa:16:3e:6f:c6:34", "network": {"id": "2c1e2031-ab59-47f5-a758-3a314aca6c1e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1100903482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "731b6700ce5b43369b40165545f04950", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1316f5aa-529f-4bac-8dd7-6076a9d43312", "external-id": "nsx-vlan-transportzone-399", "segmentation_id": 399, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43db078b-00", "ovs_interfaceid": "43db078b-0071-44a3-ab42-476acd4c56bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.567141] env[62385]: DEBUG nova.compute.manager [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 920.600441] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Releasing lock "[datastore1] devstack-image-cache_base/aa5b9007-7bfe-4917-a311-121e4cac3174" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.600441] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Processing image aa5b9007-7bfe-4917-a311-121e4cac3174 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 920.600642] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/aa5b9007-7bfe-4917-a311-121e4cac3174/aa5b9007-7bfe-4917-a311-121e4cac3174.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.600686] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquired lock "[datastore1] devstack-image-cache_base/aa5b9007-7bfe-4917-a311-121e4cac3174/aa5b9007-7bfe-4917-a311-121e4cac3174.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.600875] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 920.601589] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a491ec51-3dcc-4dab-971f-4154cb43148c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.604705] env[62385]: DEBUG nova.virt.hardware [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 920.604934] env[62385]: DEBUG nova.virt.hardware [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 920.605111] env[62385]: DEBUG nova.virt.hardware [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 920.605305] env[62385]: DEBUG nova.virt.hardware [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 920.605459] env[62385]: DEBUG nova.virt.hardware [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 920.605609] env[62385]: DEBUG nova.virt.hardware [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 920.605896] env[62385]: DEBUG nova.virt.hardware [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 920.605983] env[62385]: DEBUG nova.virt.hardware [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 920.606707] env[62385]: DEBUG nova.virt.hardware [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 920.606707] env[62385]: DEBUG nova.virt.hardware [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 920.606707] env[62385]: DEBUG nova.virt.hardware [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 920.607682] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f1b5b9a-e9ec-459f-8f27-d69a0a346810 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.617226] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb3a8c12-40ff-46b4-9652-4d5673140ed2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.632445] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 920.632599] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 920.633347] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-84ecdd17-8048-4248-8d5f-85654a5b0c04 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.639317] env[62385]: DEBUG oslo_vmware.api [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 920.639317] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52738ed9-42d8-4d3a-ca6e-b7db69999990" [ 920.639317] env[62385]: _type = "Task" [ 920.639317] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.646657] env[62385]: DEBUG oslo_vmware.api [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52738ed9-42d8-4d3a-ca6e-b7db69999990, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.715099] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c0464272-cede-4d75-8256-7a60bb9810a1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.732690] env[62385]: DEBUG nova.network.neutron [None req-53444455-1881-4ba7-b70c-4c868d7f0a6e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Updating instance_info_cache with network_info: [{"id": "f5cfcc74-0f00-411e-8ee7-e26ec0a2fe7a", "address": "fa:16:3e:18:18:ff", "network": {"id": "30c4e78b-6a64-47dc-aaaa-a210160c9f23", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1809587656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "105150c170d74e268e82ab894e9bf0d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10ff2092-e8eb-4768-ad4a-65a80560b447", "external-id": "nsx-vlan-transportzone-845", "segmentation_id": 845, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5cfcc74-0f", "ovs_interfaceid": "f5cfcc74-0f00-411e-8ee7-e26ec0a2fe7a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.988542] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9dbce20d-3a6a-44bc-85e8-6001ac82c9b8 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "b7a286a2-9a19-4ef1-8a46-6e2a04e22413" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.266s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.029477] env[62385]: DEBUG oslo_concurrency.lockutils [req-30a9c68b-faea-480e-b200-143c21a5058f req-47a8b430-815b-402b-853c-5043185c2720 service nova] Releasing lock "refresh_cache-caab700b-bd3c-497e-b40a-eb9d20dfc1e2" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.156351] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Preparing fetch location {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 921.156699] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Fetch image to [datastore1] OSTACK_IMG_22a5e0dd-c972-4959-a3bb-aa100d12f895/OSTACK_IMG_22a5e0dd-c972-4959-a3bb-aa100d12f895.vmdk {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 921.157127] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Downloading stream optimized image aa5b9007-7bfe-4917-a311-121e4cac3174 to [datastore1] OSTACK_IMG_22a5e0dd-c972-4959-a3bb-aa100d12f895/OSTACK_IMG_22a5e0dd-c972-4959-a3bb-aa100d12f895.vmdk on the data store datastore1 as vApp {{(pid=62385) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 921.157249] env[62385]: DEBUG nova.virt.vmwareapi.images [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Downloading image file data aa5b9007-7bfe-4917-a311-121e4cac3174 to the ESX as VM named 'OSTACK_IMG_22a5e0dd-c972-4959-a3bb-aa100d12f895' {{(pid=62385) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 921.198556] env[62385]: DEBUG nova.network.neutron [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Successfully updated port: f09e5aed-7121-4aae-b8e1-e6d84339486f {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 921.231260] env[62385]: DEBUG oslo_vmware.rw_handles [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 921.231260] env[62385]: value = "resgroup-9" [ 921.231260] env[62385]: _type = "ResourcePool" [ 921.231260] env[62385]: }. {{(pid=62385) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 921.231565] env[62385]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-efc1c580-edd8-4937-b054-1c1bc66072a7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.246499] env[62385]: DEBUG oslo_concurrency.lockutils [None req-53444455-1881-4ba7-b70c-4c868d7f0a6e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Releasing lock "refresh_cache-a58cb15f-e4df-4b1e-b09c-2f61dce7200f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.246742] env[62385]: DEBUG nova.objects.instance [None req-53444455-1881-4ba7-b70c-4c868d7f0a6e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lazy-loading 'migration_context' on Instance uuid a58cb15f-e4df-4b1e-b09c-2f61dce7200f {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 921.254811] env[62385]: DEBUG oslo_vmware.rw_handles [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lease: (returnval){ [ 921.254811] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]524e9e5d-3f10-41dc-ac87-5d332ea0d91c" [ 921.254811] env[62385]: _type = "HttpNfcLease" [ 921.254811] env[62385]: } obtained for vApp import into resource pool (val){ [ 921.254811] env[62385]: value = "resgroup-9" [ 921.254811] env[62385]: _type = "ResourcePool" [ 921.254811] env[62385]: }. {{(pid=62385) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 921.255102] env[62385]: DEBUG oslo_vmware.api [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the lease: (returnval){ [ 921.255102] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]524e9e5d-3f10-41dc-ac87-5d332ea0d91c" [ 921.255102] env[62385]: _type = "HttpNfcLease" [ 921.255102] env[62385]: } to be ready. {{(pid=62385) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 921.261976] env[62385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 921.261976] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]524e9e5d-3f10-41dc-ac87-5d332ea0d91c" [ 921.261976] env[62385]: _type = "HttpNfcLease" [ 921.261976] env[62385]: } is initializing. {{(pid=62385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 921.460910] env[62385]: DEBUG oslo_concurrency.lockutils [None req-45c8fa5c-7ef9-4988-b369-7b0618dbdf11 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.462162] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c68e8ca9-51ac-41ba-a715-07496c0e0bcc tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.009s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.462733] env[62385]: DEBUG nova.objects.instance [None req-c68e8ca9-51ac-41ba-a715-07496c0e0bcc tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Lazy-loading 'resources' on Instance uuid b4970e1a-2b29-44e9-b79e-cda3c32dddcd {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 921.703812] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "refresh_cache-fad3b054-cf2a-4ee3-a2a5-918a96a4a125" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.704099] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquired lock "refresh_cache-fad3b054-cf2a-4ee3-a2a5-918a96a4a125" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.704140] env[62385]: DEBUG nova.network.neutron [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 921.748989] env[62385]: DEBUG nova.objects.base [None req-53444455-1881-4ba7-b70c-4c868d7f0a6e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62385) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 921.750019] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15488c70-8f97-48e2-af55-81536646536b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.772107] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-848f0de8-c7ee-4e4d-b956-62636f028256 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.777863] env[62385]: DEBUG oslo_vmware.api [None req-53444455-1881-4ba7-b70c-4c868d7f0a6e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 921.777863] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5225ddfd-2217-a600-ade4-16e9692ce52a" [ 921.777863] env[62385]: _type = "Task" [ 921.777863] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.779531] env[62385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 921.779531] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]524e9e5d-3f10-41dc-ac87-5d332ea0d91c" [ 921.779531] env[62385]: _type = "HttpNfcLease" [ 921.779531] env[62385]: } is ready. {{(pid=62385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 921.782504] env[62385]: DEBUG oslo_vmware.rw_handles [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 921.782504] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]524e9e5d-3f10-41dc-ac87-5d332ea0d91c" [ 921.782504] env[62385]: _type = "HttpNfcLease" [ 921.782504] env[62385]: }. {{(pid=62385) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 921.783192] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4686ecb7-60e5-4540-8ba5-9e1bbfccb0a6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.790334] env[62385]: DEBUG oslo_vmware.rw_handles [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52eb380a-8ffc-c0c0-8639-0fb8bd69f17a/disk-0.vmdk from lease info. {{(pid=62385) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 921.790519] env[62385]: DEBUG oslo_vmware.rw_handles [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52eb380a-8ffc-c0c0-8639-0fb8bd69f17a/disk-0.vmdk. {{(pid=62385) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 921.794702] env[62385]: DEBUG oslo_vmware.api [None req-53444455-1881-4ba7-b70c-4c868d7f0a6e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5225ddfd-2217-a600-ade4-16e9692ce52a, 'name': SearchDatastore_Task, 'duration_secs': 0.00647} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.795896] env[62385]: DEBUG oslo_concurrency.lockutils [None req-53444455-1881-4ba7-b70c-4c868d7f0a6e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.851821] env[62385]: DEBUG nova.compute.manager [req-61ffa3a5-2370-4fb6-afa3-4eec313a3dae req-5a04374b-8d98-4f9a-9cec-c3765faa0d02 service nova] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Received event network-vif-plugged-f09e5aed-7121-4aae-b8e1-e6d84339486f {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 921.852094] env[62385]: DEBUG oslo_concurrency.lockutils [req-61ffa3a5-2370-4fb6-afa3-4eec313a3dae req-5a04374b-8d98-4f9a-9cec-c3765faa0d02 service nova] Acquiring lock "fad3b054-cf2a-4ee3-a2a5-918a96a4a125-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.852323] env[62385]: DEBUG oslo_concurrency.lockutils [req-61ffa3a5-2370-4fb6-afa3-4eec313a3dae req-5a04374b-8d98-4f9a-9cec-c3765faa0d02 service nova] Lock "fad3b054-cf2a-4ee3-a2a5-918a96a4a125-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.852508] env[62385]: DEBUG oslo_concurrency.lockutils [req-61ffa3a5-2370-4fb6-afa3-4eec313a3dae req-5a04374b-8d98-4f9a-9cec-c3765faa0d02 service nova] Lock "fad3b054-cf2a-4ee3-a2a5-918a96a4a125-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.852681] env[62385]: DEBUG nova.compute.manager [req-61ffa3a5-2370-4fb6-afa3-4eec313a3dae req-5a04374b-8d98-4f9a-9cec-c3765faa0d02 service nova] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] No waiting events found dispatching network-vif-plugged-f09e5aed-7121-4aae-b8e1-e6d84339486f {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 921.852849] env[62385]: WARNING nova.compute.manager [req-61ffa3a5-2370-4fb6-afa3-4eec313a3dae req-5a04374b-8d98-4f9a-9cec-c3765faa0d02 service nova] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Received unexpected event network-vif-plugged-f09e5aed-7121-4aae-b8e1-e6d84339486f for instance with vm_state building and task_state spawning. [ 921.853020] env[62385]: DEBUG nova.compute.manager [req-61ffa3a5-2370-4fb6-afa3-4eec313a3dae req-5a04374b-8d98-4f9a-9cec-c3765faa0d02 service nova] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Received event network-changed-f09e5aed-7121-4aae-b8e1-e6d84339486f {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 921.853183] env[62385]: DEBUG nova.compute.manager [req-61ffa3a5-2370-4fb6-afa3-4eec313a3dae req-5a04374b-8d98-4f9a-9cec-c3765faa0d02 service nova] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Refreshing instance network info cache due to event network-changed-f09e5aed-7121-4aae-b8e1-e6d84339486f. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 921.853353] env[62385]: DEBUG oslo_concurrency.lockutils [req-61ffa3a5-2370-4fb6-afa3-4eec313a3dae req-5a04374b-8d98-4f9a-9cec-c3765faa0d02 service nova] Acquiring lock "refresh_cache-fad3b054-cf2a-4ee3-a2a5-918a96a4a125" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.858848] env[62385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-510d76bb-0409-4f01-8152-bfb2d90e3d12 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.110773] env[62385]: DEBUG oslo_concurrency.lockutils [None req-cb13311b-0627-4f9e-94c6-e089812528d0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquiring lock "679150f5-ad17-45e6-9e3d-9f6e28b1aee3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.110773] env[62385]: DEBUG oslo_concurrency.lockutils [None req-cb13311b-0627-4f9e-94c6-e089812528d0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "679150f5-ad17-45e6-9e3d-9f6e28b1aee3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.110914] env[62385]: DEBUG oslo_concurrency.lockutils [None req-cb13311b-0627-4f9e-94c6-e089812528d0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquiring lock "679150f5-ad17-45e6-9e3d-9f6e28b1aee3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.111103] env[62385]: DEBUG oslo_concurrency.lockutils [None req-cb13311b-0627-4f9e-94c6-e089812528d0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "679150f5-ad17-45e6-9e3d-9f6e28b1aee3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.111301] env[62385]: DEBUG oslo_concurrency.lockutils [None req-cb13311b-0627-4f9e-94c6-e089812528d0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "679150f5-ad17-45e6-9e3d-9f6e28b1aee3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.121663] env[62385]: INFO nova.compute.manager [None req-cb13311b-0627-4f9e-94c6-e089812528d0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Terminating instance [ 922.124715] env[62385]: DEBUG nova.compute.manager [None req-cb13311b-0627-4f9e-94c6-e089812528d0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 922.124715] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-cb13311b-0627-4f9e-94c6-e089812528d0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 922.125987] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a635e6b0-18b0-4cbc-9b5b-281bff3cddcd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.136220] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb13311b-0627-4f9e-94c6-e089812528d0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 922.141186] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e318f1a5-8037-48f5-8671-55d35d3b9f2e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.148054] env[62385]: DEBUG oslo_vmware.api [None req-cb13311b-0627-4f9e-94c6-e089812528d0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 922.148054] env[62385]: value = "task-1206115" [ 922.148054] env[62385]: _type = "Task" [ 922.148054] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.157994] env[62385]: DEBUG oslo_vmware.api [None req-cb13311b-0627-4f9e-94c6-e089812528d0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206115, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.238012] env[62385]: DEBUG nova.network.neutron [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 922.287455] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19101414-1716-4fba-bc8a-bcfbc0302ffe {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.295283] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9ad2d5c-bdc2-4de2-a1f9-b21a5b5ccd83 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.332917] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf47253f-8f4a-4ed8-b463-c8899e77fe21 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.344532] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cf58c04-0cbb-4c25-83ba-d40390781af7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.360802] env[62385]: DEBUG nova.compute.provider_tree [None req-c68e8ca9-51ac-41ba-a715-07496c0e0bcc tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 922.473599] env[62385]: DEBUG nova.network.neutron [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Updating instance_info_cache with network_info: [{"id": "f09e5aed-7121-4aae-b8e1-e6d84339486f", "address": "fa:16:3e:de:ab:7e", "network": {"id": "ff21dd88-fa02-4bd6-8009-980ec6e72758", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1065789169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aea48968d4464ac0ab01c998c059107d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3ff3baee-99ce-4b51-ae98-efc6163aaab3", "external-id": "nsx-vlan-transportzone-574", "segmentation_id": 574, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf09e5aed-71", "ovs_interfaceid": "f09e5aed-7121-4aae-b8e1-e6d84339486f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.620374] env[62385]: DEBUG oslo_vmware.rw_handles [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Completed reading data from the image iterator. {{(pid=62385) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 922.620599] env[62385]: DEBUG oslo_vmware.rw_handles [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52eb380a-8ffc-c0c0-8639-0fb8bd69f17a/disk-0.vmdk. {{(pid=62385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 922.621553] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf4b8c00-db9d-42ba-bda4-44eea7da41d5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.628258] env[62385]: DEBUG oslo_vmware.rw_handles [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52eb380a-8ffc-c0c0-8639-0fb8bd69f17a/disk-0.vmdk is in state: ready. {{(pid=62385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 922.628467] env[62385]: DEBUG oslo_vmware.rw_handles [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52eb380a-8ffc-c0c0-8639-0fb8bd69f17a/disk-0.vmdk. {{(pid=62385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 922.628729] env[62385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-9152212d-578e-4dd7-bdf6-79e7dd287da2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.656877] env[62385]: DEBUG oslo_vmware.api [None req-cb13311b-0627-4f9e-94c6-e089812528d0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206115, 'name': PowerOffVM_Task, 'duration_secs': 0.205298} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.657173] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb13311b-0627-4f9e-94c6-e089812528d0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 922.657363] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-cb13311b-0627-4f9e-94c6-e089812528d0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 922.657616] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-658ad1ba-a808-48bc-b459-53f5602905e6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.723410] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-cb13311b-0627-4f9e-94c6-e089812528d0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 922.723774] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-cb13311b-0627-4f9e-94c6-e089812528d0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Deleting contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 922.723774] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb13311b-0627-4f9e-94c6-e089812528d0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Deleting the datastore file [datastore2] 679150f5-ad17-45e6-9e3d-9f6e28b1aee3 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 922.724018] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-452a7595-e244-49b8-b58c-07763465e353 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.730424] env[62385]: DEBUG oslo_vmware.api [None req-cb13311b-0627-4f9e-94c6-e089812528d0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 922.730424] env[62385]: value = "task-1206117" [ 922.730424] env[62385]: _type = "Task" [ 922.730424] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.738095] env[62385]: DEBUG oslo_vmware.api [None req-cb13311b-0627-4f9e-94c6-e089812528d0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206117, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.835783] env[62385]: DEBUG oslo_vmware.rw_handles [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52eb380a-8ffc-c0c0-8639-0fb8bd69f17a/disk-0.vmdk. {{(pid=62385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 922.836057] env[62385]: INFO nova.virt.vmwareapi.images [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Downloaded image file data aa5b9007-7bfe-4917-a311-121e4cac3174 [ 922.837164] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f2e3552-9050-4e6b-ad96-54a16927c742 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.853717] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c746998e-c7c2-4568-a528-41a1b5c6c5c6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.864489] env[62385]: DEBUG nova.scheduler.client.report [None req-c68e8ca9-51ac-41ba-a715-07496c0e0bcc tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 922.875296] env[62385]: INFO nova.virt.vmwareapi.images [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] The imported VM was unregistered [ 922.877935] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Caching image {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 922.878198] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Creating directory with path [datastore1] devstack-image-cache_base/aa5b9007-7bfe-4917-a311-121e4cac3174 {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 922.879079] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3e90d09f-2389-45b1-80f5-a408a9616fef {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.897807] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Created directory with path [datastore1] devstack-image-cache_base/aa5b9007-7bfe-4917-a311-121e4cac3174 {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 922.898033] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_22a5e0dd-c972-4959-a3bb-aa100d12f895/OSTACK_IMG_22a5e0dd-c972-4959-a3bb-aa100d12f895.vmdk to [datastore1] devstack-image-cache_base/aa5b9007-7bfe-4917-a311-121e4cac3174/aa5b9007-7bfe-4917-a311-121e4cac3174.vmdk. {{(pid=62385) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 922.898303] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-c09ca530-c689-46b2-a294-6e8faf1c0638 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.904447] env[62385]: DEBUG oslo_vmware.api [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 922.904447] env[62385]: value = "task-1206119" [ 922.904447] env[62385]: _type = "Task" [ 922.904447] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.912272] env[62385]: DEBUG oslo_vmware.api [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206119, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.976323] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Releasing lock "refresh_cache-fad3b054-cf2a-4ee3-a2a5-918a96a4a125" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.976697] env[62385]: DEBUG nova.compute.manager [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Instance network_info: |[{"id": "f09e5aed-7121-4aae-b8e1-e6d84339486f", "address": "fa:16:3e:de:ab:7e", "network": {"id": "ff21dd88-fa02-4bd6-8009-980ec6e72758", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1065789169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aea48968d4464ac0ab01c998c059107d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3ff3baee-99ce-4b51-ae98-efc6163aaab3", "external-id": "nsx-vlan-transportzone-574", "segmentation_id": 574, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf09e5aed-71", "ovs_interfaceid": "f09e5aed-7121-4aae-b8e1-e6d84339486f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 922.977069] env[62385]: DEBUG oslo_concurrency.lockutils [req-61ffa3a5-2370-4fb6-afa3-4eec313a3dae req-5a04374b-8d98-4f9a-9cec-c3765faa0d02 service nova] Acquired lock "refresh_cache-fad3b054-cf2a-4ee3-a2a5-918a96a4a125" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.977307] env[62385]: DEBUG nova.network.neutron [req-61ffa3a5-2370-4fb6-afa3-4eec313a3dae req-5a04374b-8d98-4f9a-9cec-c3765faa0d02 service nova] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Refreshing network info cache for port f09e5aed-7121-4aae-b8e1-e6d84339486f {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 922.978624] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:de:ab:7e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3ff3baee-99ce-4b51-ae98-efc6163aaab3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f09e5aed-7121-4aae-b8e1-e6d84339486f', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 922.986841] env[62385]: DEBUG oslo.service.loopingcall [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 922.987843] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 922.988092] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7eb8ca0f-63c4-47b7-aca5-e19ede9ba945 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.007539] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 923.007539] env[62385]: value = "task-1206120" [ 923.007539] env[62385]: _type = "Task" [ 923.007539] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.015282] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206120, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.243043] env[62385]: DEBUG oslo_vmware.api [None req-cb13311b-0627-4f9e-94c6-e089812528d0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206117, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.158018} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.243267] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb13311b-0627-4f9e-94c6-e089812528d0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 923.243468] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-cb13311b-0627-4f9e-94c6-e089812528d0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Deleted contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 923.243653] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-cb13311b-0627-4f9e-94c6-e089812528d0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 923.243866] env[62385]: INFO nova.compute.manager [None req-cb13311b-0627-4f9e-94c6-e089812528d0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Took 1.12 seconds to destroy the instance on the hypervisor. [ 923.244143] env[62385]: DEBUG oslo.service.loopingcall [None req-cb13311b-0627-4f9e-94c6-e089812528d0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 923.244386] env[62385]: DEBUG nova.compute.manager [-] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 923.244488] env[62385]: DEBUG nova.network.neutron [-] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 923.371390] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c68e8ca9-51ac-41ba-a715-07496c0e0bcc tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.909s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.373793] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 22.689s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.373981] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.374176] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62385) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 923.374699] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.426s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.376147] env[62385]: INFO nova.compute.claims [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 923.379575] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edc45bb8-a90e-4ef7-9b62-317e3d3d4660 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.390854] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c44a3d7f-3321-495b-8640-5928ce14a0fa {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.396482] env[62385]: INFO nova.scheduler.client.report [None req-c68e8ca9-51ac-41ba-a715-07496c0e0bcc tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Deleted allocations for instance b4970e1a-2b29-44e9-b79e-cda3c32dddcd [ 923.413521] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-161a08c6-c62a-4250-ba95-d0c2d863c587 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.426059] env[62385]: DEBUG oslo_vmware.api [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206119, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.427940] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71f6ce9e-c119-4fd8-a282-b5df79f94e90 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.464873] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179577MB free_disk=155GB free_vcpus=48 pci_devices=None {{(pid=62385) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 923.465025] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.523365] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206120, 'name': CreateVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.867953] env[62385]: DEBUG nova.network.neutron [req-61ffa3a5-2370-4fb6-afa3-4eec313a3dae req-5a04374b-8d98-4f9a-9cec-c3765faa0d02 service nova] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Updated VIF entry in instance network info cache for port f09e5aed-7121-4aae-b8e1-e6d84339486f. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 923.868855] env[62385]: DEBUG nova.network.neutron [req-61ffa3a5-2370-4fb6-afa3-4eec313a3dae req-5a04374b-8d98-4f9a-9cec-c3765faa0d02 service nova] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Updating instance_info_cache with network_info: [{"id": "f09e5aed-7121-4aae-b8e1-e6d84339486f", "address": "fa:16:3e:de:ab:7e", "network": {"id": "ff21dd88-fa02-4bd6-8009-980ec6e72758", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1065789169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aea48968d4464ac0ab01c998c059107d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3ff3baee-99ce-4b51-ae98-efc6163aaab3", "external-id": "nsx-vlan-transportzone-574", "segmentation_id": 574, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf09e5aed-71", "ovs_interfaceid": "f09e5aed-7121-4aae-b8e1-e6d84339486f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.872033] env[62385]: DEBUG nova.compute.manager [req-af49e9c3-8ddf-4f84-8c8b-fb0d980d21df req-7ec870e7-ea65-40ad-9db8-51c373de40ad service nova] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Received event network-vif-deleted-f5b51ed4-e913-4534-a2d6-33cf95ab3041 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 923.872472] env[62385]: INFO nova.compute.manager [req-af49e9c3-8ddf-4f84-8c8b-fb0d980d21df req-7ec870e7-ea65-40ad-9db8-51c373de40ad service nova] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Neutron deleted interface f5b51ed4-e913-4534-a2d6-33cf95ab3041; detaching it from the instance and deleting it from the info cache [ 923.872930] env[62385]: DEBUG nova.network.neutron [req-af49e9c3-8ddf-4f84-8c8b-fb0d980d21df req-7ec870e7-ea65-40ad-9db8-51c373de40ad service nova] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.924502] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c68e8ca9-51ac-41ba-a715-07496c0e0bcc tempest-ServerRescueTestJSON-555485555 tempest-ServerRescueTestJSON-555485555-project-member] Lock "b4970e1a-2b29-44e9-b79e-cda3c32dddcd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.517s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.933540] env[62385]: DEBUG oslo_vmware.api [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206119, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.025025] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206120, 'name': CreateVM_Task, 'duration_secs': 0.537004} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.025025] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 924.025025] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.025025] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.025025] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 924.025025] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-09691884-5dff-4e8b-9f35-4c03652d1cc1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.033900] env[62385]: DEBUG oslo_vmware.api [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 924.033900] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5254dbd4-f3cf-0a2e-4fbf-89dda836aaae" [ 924.033900] env[62385]: _type = "Task" [ 924.033900] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.045136] env[62385]: DEBUG oslo_vmware.api [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5254dbd4-f3cf-0a2e-4fbf-89dda836aaae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.130623] env[62385]: DEBUG nova.network.neutron [-] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.377103] env[62385]: DEBUG oslo_concurrency.lockutils [req-61ffa3a5-2370-4fb6-afa3-4eec313a3dae req-5a04374b-8d98-4f9a-9cec-c3765faa0d02 service nova] Releasing lock "refresh_cache-fad3b054-cf2a-4ee3-a2a5-918a96a4a125" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.377103] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e0379ad2-55ff-4f37-bec7-b63d90228f14 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.388628] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a89ab98a-9e86-4bf0-8267-ba05f9acb298 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.428631] env[62385]: DEBUG nova.compute.manager [req-af49e9c3-8ddf-4f84-8c8b-fb0d980d21df req-7ec870e7-ea65-40ad-9db8-51c373de40ad service nova] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Detach interface failed, port_id=f5b51ed4-e913-4534-a2d6-33cf95ab3041, reason: Instance 679150f5-ad17-45e6-9e3d-9f6e28b1aee3 could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 924.437022] env[62385]: DEBUG oslo_vmware.api [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206119, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.552986] env[62385]: DEBUG oslo_vmware.api [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5254dbd4-f3cf-0a2e-4fbf-89dda836aaae, 'name': SearchDatastore_Task, 'duration_secs': 0.08886} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.552986] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.552986] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 924.553378] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.553538] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.555076] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 924.555076] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-62caa60d-d4a2-42fb-bc73-8c5dbcaccb97 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.570625] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 924.570798] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 924.573179] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4eb3bb10-e1ad-488a-8fba-eef2efd50261 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.580243] env[62385]: DEBUG oslo_vmware.api [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 924.580243] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ef0200-85c1-33a8-add0-1e516d268b93" [ 924.580243] env[62385]: _type = "Task" [ 924.580243] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.594870] env[62385]: DEBUG oslo_vmware.api [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ef0200-85c1-33a8-add0-1e516d268b93, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.632602] env[62385]: INFO nova.compute.manager [-] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Took 1.39 seconds to deallocate network for instance. [ 924.746276] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2481d37c-e203-40d6-8b21-fe77f9614951 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.757119] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c057c82-e682-490b-a90b-77c3657481e3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.791488] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24a909ca-c823-441a-850a-14ab8647b25e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.802210] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92f497aa-042b-4823-bc29-d06b9d9995f7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.819671] env[62385]: DEBUG nova.compute.provider_tree [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Updating inventory in ProviderTree for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 924.925073] env[62385]: DEBUG oslo_vmware.api [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206119, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.104252] env[62385]: DEBUG oslo_vmware.api [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ef0200-85c1-33a8-add0-1e516d268b93, 'name': SearchDatastore_Task, 'duration_secs': 0.092487} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.105128] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cdbe9cdd-168c-4b9f-a2fa-76fe4d55e981 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.115893] env[62385]: DEBUG oslo_vmware.api [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 925.115893] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5286b57a-e255-e34d-8ad9-64355ea74977" [ 925.115893] env[62385]: _type = "Task" [ 925.115893] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.127902] env[62385]: DEBUG oslo_vmware.api [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5286b57a-e255-e34d-8ad9-64355ea74977, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.142425] env[62385]: DEBUG oslo_concurrency.lockutils [None req-cb13311b-0627-4f9e-94c6-e089812528d0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.355015] env[62385]: DEBUG nova.scheduler.client.report [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Updated inventory for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a with generation 108 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 925.355334] env[62385]: DEBUG nova.compute.provider_tree [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Updating resource provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a generation from 108 to 109 during operation: update_inventory {{(pid=62385) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 925.355528] env[62385]: DEBUG nova.compute.provider_tree [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Updating inventory in ProviderTree for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 925.423584] env[62385]: DEBUG oslo_vmware.api [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206119, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.627485] env[62385]: DEBUG oslo_vmware.api [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5286b57a-e255-e34d-8ad9-64355ea74977, 'name': SearchDatastore_Task, 'duration_secs': 0.07993} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.627773] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.628445] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] fad3b054-cf2a-4ee3-a2a5-918a96a4a125/fad3b054-cf2a-4ee3-a2a5-918a96a4a125.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 925.628763] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-aea86fad-7003-4651-a86b-12f63a15e3f5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.637304] env[62385]: DEBUG oslo_vmware.api [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 925.637304] env[62385]: value = "task-1206121" [ 925.637304] env[62385]: _type = "Task" [ 925.637304] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.645166] env[62385]: DEBUG oslo_vmware.api [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206121, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.860888] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.486s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.861448] env[62385]: DEBUG nova.compute.manager [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 925.864055] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6287d190-3499-494d-8c3c-9507cccb256b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.950s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.864289] env[62385]: DEBUG nova.objects.instance [None req-6287d190-3499-494d-8c3c-9507cccb256b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lazy-loading 'resources' on Instance uuid d12ffea4-2275-4974-a5c6-b7fa365865b4 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 925.923253] env[62385]: DEBUG oslo_vmware.api [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206119, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.640228} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.923557] env[62385]: INFO nova.virt.vmwareapi.ds_util [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_22a5e0dd-c972-4959-a3bb-aa100d12f895/OSTACK_IMG_22a5e0dd-c972-4959-a3bb-aa100d12f895.vmdk to [datastore1] devstack-image-cache_base/aa5b9007-7bfe-4917-a311-121e4cac3174/aa5b9007-7bfe-4917-a311-121e4cac3174.vmdk. [ 925.923758] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Cleaning up location [datastore1] OSTACK_IMG_22a5e0dd-c972-4959-a3bb-aa100d12f895 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 925.923930] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_22a5e0dd-c972-4959-a3bb-aa100d12f895 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 925.924225] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d3a4f15d-4dfc-41ad-b8d3-29139848293c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.930812] env[62385]: DEBUG oslo_vmware.api [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 925.930812] env[62385]: value = "task-1206122" [ 925.930812] env[62385]: _type = "Task" [ 925.930812] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.938167] env[62385]: DEBUG oslo_vmware.api [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206122, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.152207] env[62385]: DEBUG oslo_vmware.api [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206121, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.369730] env[62385]: DEBUG nova.compute.utils [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 926.374165] env[62385]: DEBUG nova.compute.manager [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 926.374165] env[62385]: DEBUG nova.network.neutron [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 926.439428] env[62385]: DEBUG nova.policy [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '923d29b88d064a1da592e1175fd8fe4b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd68661a0197d4e3cacc06ea433ac706e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 926.444537] env[62385]: DEBUG oslo_vmware.api [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206122, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.056205} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.444785] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 926.444958] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Releasing lock "[datastore1] devstack-image-cache_base/aa5b9007-7bfe-4917-a311-121e4cac3174/aa5b9007-7bfe-4917-a311-121e4cac3174.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.445232] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/aa5b9007-7bfe-4917-a311-121e4cac3174/aa5b9007-7bfe-4917-a311-121e4cac3174.vmdk to [datastore1] caab700b-bd3c-497e-b40a-eb9d20dfc1e2/caab700b-bd3c-497e-b40a-eb9d20dfc1e2.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 926.445563] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-da16c308-430a-4e47-bb0b-4576cf83e6c0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.454869] env[62385]: DEBUG oslo_vmware.api [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 926.454869] env[62385]: value = "task-1206123" [ 926.454869] env[62385]: _type = "Task" [ 926.454869] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.466076] env[62385]: DEBUG oslo_vmware.api [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206123, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.649995] env[62385]: DEBUG oslo_vmware.api [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206121, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.595679} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.650292] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] fad3b054-cf2a-4ee3-a2a5-918a96a4a125/fad3b054-cf2a-4ee3-a2a5-918a96a4a125.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 926.650592] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 926.650848] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-28f26c2c-91dd-4b39-a215-5a06e559dcea {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.665346] env[62385]: DEBUG oslo_vmware.api [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 926.665346] env[62385]: value = "task-1206124" [ 926.665346] env[62385]: _type = "Task" [ 926.665346] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.677107] env[62385]: DEBUG oslo_vmware.api [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206124, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.692082] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eda4dcb3-e7db-4d04-82a0-f506ffc1998d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.701807] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e499a9b9-6cb1-4590-977c-4a75a871bc5d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.736273] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4750306c-236f-4dd2-b87d-3ddb770c2c77 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.745311] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-530265db-0b0e-41c7-b0a6-315040b21821 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.761446] env[62385]: DEBUG nova.compute.provider_tree [None req-6287d190-3499-494d-8c3c-9507cccb256b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 926.876065] env[62385]: DEBUG nova.compute.manager [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 926.879554] env[62385]: DEBUG nova.network.neutron [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Successfully created port: 89da5ce4-4773-4c11-91a4-870768a5b165 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 926.967288] env[62385]: DEBUG oslo_vmware.api [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206123, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.184400] env[62385]: DEBUG oslo_vmware.api [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206124, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075779} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.184734] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 927.185585] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69aa17c5-651a-41a0-bac9-0902eff8e3dc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.210759] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Reconfiguring VM instance instance-00000059 to attach disk [datastore1] fad3b054-cf2a-4ee3-a2a5-918a96a4a125/fad3b054-cf2a-4ee3-a2a5-918a96a4a125.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 927.211406] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6f0300b6-bda6-4637-8d2b-63ad858717c2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.234554] env[62385]: DEBUG oslo_vmware.api [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 927.234554] env[62385]: value = "task-1206125" [ 927.234554] env[62385]: _type = "Task" [ 927.234554] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.235399] env[62385]: DEBUG nova.network.neutron [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Successfully created port: e2a123f6-199e-4e54-bd44-8b64435eebca {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 927.248173] env[62385]: DEBUG oslo_vmware.api [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206125, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.265018] env[62385]: DEBUG nova.scheduler.client.report [None req-6287d190-3499-494d-8c3c-9507cccb256b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 927.469686] env[62385]: DEBUG oslo_vmware.api [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206123, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.749971] env[62385]: DEBUG oslo_vmware.api [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206125, 'name': ReconfigVM_Task, 'duration_secs': 0.324611} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.750315] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Reconfigured VM instance instance-00000059 to attach disk [datastore1] fad3b054-cf2a-4ee3-a2a5-918a96a4a125/fad3b054-cf2a-4ee3-a2a5-918a96a4a125.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 927.751071] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a03bd33b-9ae9-4cad-a943-2052de840466 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.760730] env[62385]: DEBUG oslo_vmware.api [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 927.760730] env[62385]: value = "task-1206126" [ 927.760730] env[62385]: _type = "Task" [ 927.760730] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.771725] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6287d190-3499-494d-8c3c-9507cccb256b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.908s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.773971] env[62385]: DEBUG oslo_vmware.api [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206126, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.774503] env[62385]: DEBUG oslo_concurrency.lockutils [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 24.064s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.774715] env[62385]: DEBUG nova.objects.instance [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62385) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 927.806863] env[62385]: INFO nova.scheduler.client.report [None req-6287d190-3499-494d-8c3c-9507cccb256b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Deleted allocations for instance d12ffea4-2275-4974-a5c6-b7fa365865b4 [ 927.885886] env[62385]: DEBUG nova.compute.manager [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 927.915016] env[62385]: DEBUG nova.virt.hardware [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 927.915324] env[62385]: DEBUG nova.virt.hardware [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 927.915499] env[62385]: DEBUG nova.virt.hardware [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 927.915685] env[62385]: DEBUG nova.virt.hardware [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 927.915838] env[62385]: DEBUG nova.virt.hardware [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 927.915991] env[62385]: DEBUG nova.virt.hardware [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 927.916269] env[62385]: DEBUG nova.virt.hardware [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 927.916473] env[62385]: DEBUG nova.virt.hardware [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 927.916659] env[62385]: DEBUG nova.virt.hardware [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 927.916830] env[62385]: DEBUG nova.virt.hardware [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 927.917034] env[62385]: DEBUG nova.virt.hardware [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 927.918106] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f886f198-cb2b-46b1-bde6-10f831fa4d86 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.928482] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95280554-a1bd-4d82-b44e-6dccd49122c5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.968182] env[62385]: DEBUG oslo_vmware.api [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206123, 'name': CopyVirtualDisk_Task} progress is 66%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.273887] env[62385]: DEBUG oslo_vmware.api [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206126, 'name': Rename_Task, 'duration_secs': 0.157009} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.274509] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 928.274845] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a429bc0b-8cc1-4d22-a492-999c1edec0c4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.284393] env[62385]: DEBUG oslo_vmware.api [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 928.284393] env[62385]: value = "task-1206127" [ 928.284393] env[62385]: _type = "Task" [ 928.284393] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.295670] env[62385]: DEBUG oslo_vmware.api [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206127, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.315026] env[62385]: DEBUG oslo_concurrency.lockutils [None req-6287d190-3499-494d-8c3c-9507cccb256b tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "d12ffea4-2275-4974-a5c6-b7fa365865b4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.833s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.474253] env[62385]: DEBUG oslo_vmware.api [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206123, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.757589] env[62385]: DEBUG nova.compute.manager [req-dff0ed36-c7fb-488f-b5d4-c77fee143187 req-4b70d36e-0342-4e43-a784-776156ad9cb7 service nova] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Received event network-vif-plugged-89da5ce4-4773-4c11-91a4-870768a5b165 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 928.757589] env[62385]: DEBUG oslo_concurrency.lockutils [req-dff0ed36-c7fb-488f-b5d4-c77fee143187 req-4b70d36e-0342-4e43-a784-776156ad9cb7 service nova] Acquiring lock "75cf7907-d9e3-4f54-90c9-f8d714e1df40-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.757589] env[62385]: DEBUG oslo_concurrency.lockutils [req-dff0ed36-c7fb-488f-b5d4-c77fee143187 req-4b70d36e-0342-4e43-a784-776156ad9cb7 service nova] Lock "75cf7907-d9e3-4f54-90c9-f8d714e1df40-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.757589] env[62385]: DEBUG oslo_concurrency.lockutils [req-dff0ed36-c7fb-488f-b5d4-c77fee143187 req-4b70d36e-0342-4e43-a784-776156ad9cb7 service nova] Lock "75cf7907-d9e3-4f54-90c9-f8d714e1df40-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.757589] env[62385]: DEBUG nova.compute.manager [req-dff0ed36-c7fb-488f-b5d4-c77fee143187 req-4b70d36e-0342-4e43-a784-776156ad9cb7 service nova] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] No waiting events found dispatching network-vif-plugged-89da5ce4-4773-4c11-91a4-870768a5b165 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 928.757589] env[62385]: WARNING nova.compute.manager [req-dff0ed36-c7fb-488f-b5d4-c77fee143187 req-4b70d36e-0342-4e43-a784-776156ad9cb7 service nova] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Received unexpected event network-vif-plugged-89da5ce4-4773-4c11-91a4-870768a5b165 for instance with vm_state building and task_state spawning. [ 928.787024] env[62385]: DEBUG oslo_concurrency.lockutils [None req-937e1824-994e-4975-a9f1-f116c79c4fd2 tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.787024] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67f292fc-5c39-4e07-ad8d-7dc61472e3a5 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.706s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.787024] env[62385]: DEBUG nova.objects.instance [None req-67f292fc-5c39-4e07-ad8d-7dc61472e3a5 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Lazy-loading 'resources' on Instance uuid e4089e71-eea1-4e16-b90c-97966333f2bc {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 928.801145] env[62385]: DEBUG oslo_vmware.api [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206127, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.970313] env[62385]: DEBUG oslo_vmware.api [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206123, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.386408} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.971445] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/aa5b9007-7bfe-4917-a311-121e4cac3174/aa5b9007-7bfe-4917-a311-121e4cac3174.vmdk to [datastore1] caab700b-bd3c-497e-b40a-eb9d20dfc1e2/caab700b-bd3c-497e-b40a-eb9d20dfc1e2.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 928.972688] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3432f90e-ca89-49a2-aebb-257b62698402 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.996184] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] caab700b-bd3c-497e-b40a-eb9d20dfc1e2/caab700b-bd3c-497e-b40a-eb9d20dfc1e2.vmdk or device None with type streamOptimized {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 928.996536] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8be08f26-2963-4c9e-a5be-06333c738bb7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.016518] env[62385]: DEBUG oslo_vmware.api [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 929.016518] env[62385]: value = "task-1206128" [ 929.016518] env[62385]: _type = "Task" [ 929.016518] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.024981] env[62385]: DEBUG oslo_vmware.api [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206128, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.300517] env[62385]: DEBUG oslo_vmware.api [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206127, 'name': PowerOnVM_Task, 'duration_secs': 0.631506} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.300799] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 929.301249] env[62385]: INFO nova.compute.manager [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Took 8.73 seconds to spawn the instance on the hypervisor. [ 929.301249] env[62385]: DEBUG nova.compute.manager [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 929.302162] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8acb2595-4b18-4a94-bb69-dc65cadbc007 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.421175] env[62385]: DEBUG nova.network.neutron [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Successfully updated port: 89da5ce4-4773-4c11-91a4-870768a5b165 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 929.451848] env[62385]: DEBUG nova.compute.manager [req-805df04f-d268-4ca2-9d42-d9ba07afed4b req-e4ddfa4d-3258-40c0-9c32-7eafcc03752d service nova] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Received event network-changed-89da5ce4-4773-4c11-91a4-870768a5b165 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 929.452147] env[62385]: DEBUG nova.compute.manager [req-805df04f-d268-4ca2-9d42-d9ba07afed4b req-e4ddfa4d-3258-40c0-9c32-7eafcc03752d service nova] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Refreshing instance network info cache due to event network-changed-89da5ce4-4773-4c11-91a4-870768a5b165. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 929.452248] env[62385]: DEBUG oslo_concurrency.lockutils [req-805df04f-d268-4ca2-9d42-d9ba07afed4b req-e4ddfa4d-3258-40c0-9c32-7eafcc03752d service nova] Acquiring lock "refresh_cache-75cf7907-d9e3-4f54-90c9-f8d714e1df40" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.452467] env[62385]: DEBUG oslo_concurrency.lockutils [req-805df04f-d268-4ca2-9d42-d9ba07afed4b req-e4ddfa4d-3258-40c0-9c32-7eafcc03752d service nova] Acquired lock "refresh_cache-75cf7907-d9e3-4f54-90c9-f8d714e1df40" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.452533] env[62385]: DEBUG nova.network.neutron [req-805df04f-d268-4ca2-9d42-d9ba07afed4b req-e4ddfa4d-3258-40c0-9c32-7eafcc03752d service nova] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Refreshing network info cache for port 89da5ce4-4773-4c11-91a4-870768a5b165 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 929.530962] env[62385]: DEBUG oslo_vmware.api [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206128, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.562482] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-925b487c-e9f0-4111-82af-edf614669e59 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.570051] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac9ae686-2ced-464a-b412-525f21489c50 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.599714] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbd5023c-395d-409c-8edb-5f3491ec11a8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.606900] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e5ed40b-a529-4e63-ad27-69fa1649f235 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.619881] env[62385]: DEBUG nova.compute.provider_tree [None req-67f292fc-5c39-4e07-ad8d-7dc61472e3a5 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 929.822783] env[62385]: INFO nova.compute.manager [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Took 34.98 seconds to build instance. [ 930.029859] env[62385]: DEBUG oslo_vmware.api [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206128, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.047610] env[62385]: DEBUG nova.network.neutron [req-805df04f-d268-4ca2-9d42-d9ba07afed4b req-e4ddfa4d-3258-40c0-9c32-7eafcc03752d service nova] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 930.123084] env[62385]: DEBUG nova.scheduler.client.report [None req-67f292fc-5c39-4e07-ad8d-7dc61472e3a5 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 930.324415] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ed2cc41-2ddb-4bcc-b10f-266b208d4a7b tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "fad3b054-cf2a-4ee3-a2a5-918a96a4a125" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.490s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.338529] env[62385]: DEBUG nova.network.neutron [req-805df04f-d268-4ca2-9d42-d9ba07afed4b req-e4ddfa4d-3258-40c0-9c32-7eafcc03752d service nova] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.535018] env[62385]: DEBUG oslo_vmware.api [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206128, 'name': ReconfigVM_Task, 'duration_secs': 1.276231} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.535018] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Reconfigured VM instance instance-00000058 to attach disk [datastore1] caab700b-bd3c-497e-b40a-eb9d20dfc1e2/caab700b-bd3c-497e-b40a-eb9d20dfc1e2.vmdk or device None with type streamOptimized {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 930.535018] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fb356911-7b15-4aca-9799-5eacaad67a07 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.541204] env[62385]: DEBUG oslo_vmware.api [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 930.541204] env[62385]: value = "task-1206129" [ 930.541204] env[62385]: _type = "Task" [ 930.541204] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.549653] env[62385]: DEBUG oslo_vmware.api [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206129, 'name': Rename_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.630170] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67f292fc-5c39-4e07-ad8d-7dc61472e3a5 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.844s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.632531] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ada74c47-0cb2-4e31-bee2-83de1a49589a tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.474s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.632765] env[62385]: DEBUG nova.objects.instance [None req-ada74c47-0cb2-4e31-bee2-83de1a49589a tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Lazy-loading 'resources' on Instance uuid 2bcd9457-8a73-4e7a-b778-d52c468b3aae {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 930.653139] env[62385]: INFO nova.scheduler.client.report [None req-67f292fc-5c39-4e07-ad8d-7dc61472e3a5 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Deleted allocations for instance e4089e71-eea1-4e16-b90c-97966333f2bc [ 930.841221] env[62385]: DEBUG oslo_concurrency.lockutils [req-805df04f-d268-4ca2-9d42-d9ba07afed4b req-e4ddfa4d-3258-40c0-9c32-7eafcc03752d service nova] Releasing lock "refresh_cache-75cf7907-d9e3-4f54-90c9-f8d714e1df40" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.018454] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85272c60-14b4-41ec-9409-1104e80731e4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.025451] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a3883fb0-9fe3-4b99-97f5-f95d16cbbde3 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Suspending the VM {{(pid=62385) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 931.025701] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-96352459-fd6c-462e-83d4-2a01fd63c527 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.032829] env[62385]: DEBUG oslo_vmware.api [None req-a3883fb0-9fe3-4b99-97f5-f95d16cbbde3 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 931.032829] env[62385]: value = "task-1206130" [ 931.032829] env[62385]: _type = "Task" [ 931.032829] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.040561] env[62385]: DEBUG oslo_vmware.api [None req-a3883fb0-9fe3-4b99-97f5-f95d16cbbde3 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206130, 'name': SuspendVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.050077] env[62385]: DEBUG oslo_vmware.api [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206129, 'name': Rename_Task, 'duration_secs': 0.141146} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.050487] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 931.050575] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3db8840a-dd8f-4502-ae7b-d9abf5be1176 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.057374] env[62385]: DEBUG oslo_vmware.api [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 931.057374] env[62385]: value = "task-1206131" [ 931.057374] env[62385]: _type = "Task" [ 931.057374] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.070021] env[62385]: DEBUG oslo_vmware.api [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206131, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.161779] env[62385]: DEBUG oslo_concurrency.lockutils [None req-67f292fc-5c39-4e07-ad8d-7dc61472e3a5 tempest-ListImageFiltersTestJSON-1494384867 tempest-ListImageFiltersTestJSON-1494384867-project-member] Lock "e4089e71-eea1-4e16-b90c-97966333f2bc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.971s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.384330] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-827e3951-c6a7-47c6-b69d-48af35182e44 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.393291] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dffc1e75-caf7-4ef1-8171-ee305317db41 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.428264] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6132ffb3-55ad-4f51-aa80-8a2957582b3f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.437200] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c806857-415d-4d30-a182-a574a1cedfc6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.451125] env[62385]: DEBUG nova.compute.provider_tree [None req-ada74c47-0cb2-4e31-bee2-83de1a49589a tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 931.543460] env[62385]: DEBUG oslo_vmware.api [None req-a3883fb0-9fe3-4b99-97f5-f95d16cbbde3 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206130, 'name': SuspendVM_Task} progress is 66%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.567983] env[62385]: DEBUG oslo_vmware.api [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206131, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.661097] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "5daf0a99-0c2a-40d8-afc7-1998e21e32d8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.661348] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "5daf0a99-0c2a-40d8-afc7-1998e21e32d8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.954843] env[62385]: DEBUG nova.scheduler.client.report [None req-ada74c47-0cb2-4e31-bee2-83de1a49589a tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 932.005148] env[62385]: DEBUG nova.compute.manager [req-9c714c99-9e04-4bcf-b42f-8fc7e548e5b9 req-1823e9ed-056b-45be-836c-05f9795fa12b service nova] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Received event network-vif-plugged-e2a123f6-199e-4e54-bd44-8b64435eebca {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 932.005425] env[62385]: DEBUG oslo_concurrency.lockutils [req-9c714c99-9e04-4bcf-b42f-8fc7e548e5b9 req-1823e9ed-056b-45be-836c-05f9795fa12b service nova] Acquiring lock "75cf7907-d9e3-4f54-90c9-f8d714e1df40-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.005596] env[62385]: DEBUG oslo_concurrency.lockutils [req-9c714c99-9e04-4bcf-b42f-8fc7e548e5b9 req-1823e9ed-056b-45be-836c-05f9795fa12b service nova] Lock "75cf7907-d9e3-4f54-90c9-f8d714e1df40-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.005781] env[62385]: DEBUG oslo_concurrency.lockutils [req-9c714c99-9e04-4bcf-b42f-8fc7e548e5b9 req-1823e9ed-056b-45be-836c-05f9795fa12b service nova] Lock "75cf7907-d9e3-4f54-90c9-f8d714e1df40-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.005952] env[62385]: DEBUG nova.compute.manager [req-9c714c99-9e04-4bcf-b42f-8fc7e548e5b9 req-1823e9ed-056b-45be-836c-05f9795fa12b service nova] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] No waiting events found dispatching network-vif-plugged-e2a123f6-199e-4e54-bd44-8b64435eebca {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 932.006269] env[62385]: WARNING nova.compute.manager [req-9c714c99-9e04-4bcf-b42f-8fc7e548e5b9 req-1823e9ed-056b-45be-836c-05f9795fa12b service nova] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Received unexpected event network-vif-plugged-e2a123f6-199e-4e54-bd44-8b64435eebca for instance with vm_state building and task_state spawning. [ 932.046166] env[62385]: DEBUG oslo_vmware.api [None req-a3883fb0-9fe3-4b99-97f5-f95d16cbbde3 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206130, 'name': SuspendVM_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.066555] env[62385]: DEBUG oslo_vmware.api [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206131, 'name': PowerOnVM_Task, 'duration_secs': 0.542856} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.066824] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 932.067090] env[62385]: INFO nova.compute.manager [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Took 14.99 seconds to spawn the instance on the hypervisor. [ 932.067314] env[62385]: DEBUG nova.compute.manager [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 932.068092] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a78734e-1b79-40d1-a1dc-b843ee4cd373 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.109439] env[62385]: DEBUG nova.network.neutron [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Successfully updated port: e2a123f6-199e-4e54-bd44-8b64435eebca {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 932.163925] env[62385]: DEBUG nova.compute.manager [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 932.460047] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ada74c47-0cb2-4e31-bee2-83de1a49589a tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.827s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.462494] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.574s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.464049] env[62385]: INFO nova.compute.claims [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 932.486210] env[62385]: INFO nova.scheduler.client.report [None req-ada74c47-0cb2-4e31-bee2-83de1a49589a tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Deleted allocations for instance 2bcd9457-8a73-4e7a-b778-d52c468b3aae [ 932.544834] env[62385]: DEBUG oslo_vmware.api [None req-a3883fb0-9fe3-4b99-97f5-f95d16cbbde3 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206130, 'name': SuspendVM_Task, 'duration_secs': 1.046006} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.547147] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a3883fb0-9fe3-4b99-97f5-f95d16cbbde3 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Suspended the VM {{(pid=62385) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 932.547391] env[62385]: DEBUG nova.compute.manager [None req-a3883fb0-9fe3-4b99-97f5-f95d16cbbde3 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 932.548190] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-695e3817-617f-4f8c-b04f-bd27151e840e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.586468] env[62385]: INFO nova.compute.manager [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Took 37.77 seconds to build instance. [ 932.611765] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Acquiring lock "refresh_cache-75cf7907-d9e3-4f54-90c9-f8d714e1df40" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.611926] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Acquired lock "refresh_cache-75cf7907-d9e3-4f54-90c9-f8d714e1df40" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.612091] env[62385]: DEBUG nova.network.neutron [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 932.682950] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.994446] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ada74c47-0cb2-4e31-bee2-83de1a49589a tempest-ServerShowV254Test-2107323780 tempest-ServerShowV254Test-2107323780-project-member] Lock "2bcd9457-8a73-4e7a-b778-d52c468b3aae" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.645s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.088217] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ec27b628-11d7-4948-a444-7a281fe62904 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "caab700b-bd3c-497e-b40a-eb9d20dfc1e2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.282s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.180311] env[62385]: DEBUG nova.network.neutron [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 933.537022] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e75e6879-4fef-42a9-ba97-4fe18524ff8a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquiring lock "caab700b-bd3c-497e-b40a-eb9d20dfc1e2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.539068] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e75e6879-4fef-42a9-ba97-4fe18524ff8a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "caab700b-bd3c-497e-b40a-eb9d20dfc1e2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.002s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.539409] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e75e6879-4fef-42a9-ba97-4fe18524ff8a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquiring lock "caab700b-bd3c-497e-b40a-eb9d20dfc1e2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.540106] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e75e6879-4fef-42a9-ba97-4fe18524ff8a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "caab700b-bd3c-497e-b40a-eb9d20dfc1e2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.540567] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e75e6879-4fef-42a9-ba97-4fe18524ff8a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "caab700b-bd3c-497e-b40a-eb9d20dfc1e2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.543269] env[62385]: INFO nova.compute.manager [None req-e75e6879-4fef-42a9-ba97-4fe18524ff8a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Terminating instance [ 933.550926] env[62385]: DEBUG nova.compute.manager [None req-e75e6879-4fef-42a9-ba97-4fe18524ff8a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 933.551156] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-e75e6879-4fef-42a9-ba97-4fe18524ff8a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 933.552038] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06a85bf6-079e-4e9f-9149-4f946312f44b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.560345] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e75e6879-4fef-42a9-ba97-4fe18524ff8a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 933.560611] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-09c4298a-d349-42d7-b59c-03808edfa310 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.569840] env[62385]: DEBUG oslo_vmware.api [None req-e75e6879-4fef-42a9-ba97-4fe18524ff8a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 933.569840] env[62385]: value = "task-1206132" [ 933.569840] env[62385]: _type = "Task" [ 933.569840] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.579213] env[62385]: DEBUG oslo_vmware.api [None req-e75e6879-4fef-42a9-ba97-4fe18524ff8a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206132, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.708440] env[62385]: DEBUG nova.network.neutron [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Updating instance_info_cache with network_info: [{"id": "89da5ce4-4773-4c11-91a4-870768a5b165", "address": "fa:16:3e:d3:20:43", "network": {"id": "58340331-dbdb-4903-bab1-f17d05e248d3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1767092846", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.56", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d68661a0197d4e3cacc06ea433ac706e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5116f690-f825-4fee-8a47-42b073e716c5", "external-id": "nsx-vlan-transportzone-692", "segmentation_id": 692, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap89da5ce4-47", "ovs_interfaceid": "89da5ce4-4773-4c11-91a4-870768a5b165", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e2a123f6-199e-4e54-bd44-8b64435eebca", "address": "fa:16:3e:6d:a5:74", "network": {"id": "bba82320-3049-4f4a-a76e-52acce8b75ee", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-310743232", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.149", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "d68661a0197d4e3cacc06ea433ac706e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a64108f9-df0a-4feb-bbb5-97f5841c356c", "external-id": "nsx-vlan-transportzone-67", "segmentation_id": 67, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape2a123f6-19", "ovs_interfaceid": "e2a123f6-199e-4e54-bd44-8b64435eebca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.790445] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f3c51ad-0209-4ef1-aca2-ba5faf7bf857 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.799440] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1f3804f-b51c-4f00-b63a-12cf0a3d0efe {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.837786] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f7e0854-92de-49e1-b897-b712a65fe53a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.846582] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8065d7d8-25e5-4a58-b20c-9d7bfdf2e90a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.862237] env[62385]: DEBUG nova.compute.provider_tree [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 933.891234] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0d3ddba5-96d8-4d1f-a0cb-bb6183881104 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "fad3b054-cf2a-4ee3-a2a5-918a96a4a125" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.891494] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0d3ddba5-96d8-4d1f-a0cb-bb6183881104 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "fad3b054-cf2a-4ee3-a2a5-918a96a4a125" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.891706] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0d3ddba5-96d8-4d1f-a0cb-bb6183881104 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "fad3b054-cf2a-4ee3-a2a5-918a96a4a125-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.891893] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0d3ddba5-96d8-4d1f-a0cb-bb6183881104 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "fad3b054-cf2a-4ee3-a2a5-918a96a4a125-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.892080] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0d3ddba5-96d8-4d1f-a0cb-bb6183881104 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "fad3b054-cf2a-4ee3-a2a5-918a96a4a125-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.894204] env[62385]: INFO nova.compute.manager [None req-0d3ddba5-96d8-4d1f-a0cb-bb6183881104 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Terminating instance [ 933.896147] env[62385]: DEBUG nova.compute.manager [None req-0d3ddba5-96d8-4d1f-a0cb-bb6183881104 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 933.896258] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0d3ddba5-96d8-4d1f-a0cb-bb6183881104 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 933.897056] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f55df7b9-0604-4a69-ab5f-03ab2bd85816 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.905437] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0d3ddba5-96d8-4d1f-a0cb-bb6183881104 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 933.905910] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8eb6bfb5-dcf1-4f8c-91d2-ebe1e331f506 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.974205] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0d3ddba5-96d8-4d1f-a0cb-bb6183881104 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 933.974472] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0d3ddba5-96d8-4d1f-a0cb-bb6183881104 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 933.974534] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d3ddba5-96d8-4d1f-a0cb-bb6183881104 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Deleting the datastore file [datastore1] fad3b054-cf2a-4ee3-a2a5-918a96a4a125 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 933.974792] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-04dec2ed-f47b-4f60-8254-c0037ca6e33d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.981440] env[62385]: DEBUG oslo_vmware.api [None req-0d3ddba5-96d8-4d1f-a0cb-bb6183881104 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 933.981440] env[62385]: value = "task-1206134" [ 933.981440] env[62385]: _type = "Task" [ 933.981440] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.988941] env[62385]: DEBUG oslo_vmware.api [None req-0d3ddba5-96d8-4d1f-a0cb-bb6183881104 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206134, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.033299] env[62385]: DEBUG nova.compute.manager [req-fb047e78-3cf4-4d01-bbe4-83f54fa18a7d req-6d19663f-094a-4a35-972f-0a2bcc764710 service nova] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Received event network-changed-e2a123f6-199e-4e54-bd44-8b64435eebca {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 934.033497] env[62385]: DEBUG nova.compute.manager [req-fb047e78-3cf4-4d01-bbe4-83f54fa18a7d req-6d19663f-094a-4a35-972f-0a2bcc764710 service nova] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Refreshing instance network info cache due to event network-changed-e2a123f6-199e-4e54-bd44-8b64435eebca. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 934.033718] env[62385]: DEBUG oslo_concurrency.lockutils [req-fb047e78-3cf4-4d01-bbe4-83f54fa18a7d req-6d19663f-094a-4a35-972f-0a2bcc764710 service nova] Acquiring lock "refresh_cache-75cf7907-d9e3-4f54-90c9-f8d714e1df40" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.079309] env[62385]: DEBUG oslo_vmware.api [None req-e75e6879-4fef-42a9-ba97-4fe18524ff8a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206132, 'name': PowerOffVM_Task, 'duration_secs': 0.204047} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.079764] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e75e6879-4fef-42a9-ba97-4fe18524ff8a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 934.079967] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-e75e6879-4fef-42a9-ba97-4fe18524ff8a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 934.080260] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-89bebacd-5d0a-43fa-9da2-651cbe03b9f4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.210100] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Releasing lock "refresh_cache-75cf7907-d9e3-4f54-90c9-f8d714e1df40" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.210503] env[62385]: DEBUG nova.compute.manager [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Instance network_info: |[{"id": "89da5ce4-4773-4c11-91a4-870768a5b165", "address": "fa:16:3e:d3:20:43", "network": {"id": "58340331-dbdb-4903-bab1-f17d05e248d3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1767092846", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.56", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d68661a0197d4e3cacc06ea433ac706e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5116f690-f825-4fee-8a47-42b073e716c5", "external-id": "nsx-vlan-transportzone-692", "segmentation_id": 692, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap89da5ce4-47", "ovs_interfaceid": "89da5ce4-4773-4c11-91a4-870768a5b165", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e2a123f6-199e-4e54-bd44-8b64435eebca", "address": "fa:16:3e:6d:a5:74", "network": {"id": "bba82320-3049-4f4a-a76e-52acce8b75ee", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-310743232", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.149", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "d68661a0197d4e3cacc06ea433ac706e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a64108f9-df0a-4feb-bbb5-97f5841c356c", "external-id": "nsx-vlan-transportzone-67", "segmentation_id": 67, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape2a123f6-19", "ovs_interfaceid": "e2a123f6-199e-4e54-bd44-8b64435eebca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 934.210839] env[62385]: DEBUG oslo_concurrency.lockutils [req-fb047e78-3cf4-4d01-bbe4-83f54fa18a7d req-6d19663f-094a-4a35-972f-0a2bcc764710 service nova] Acquired lock "refresh_cache-75cf7907-d9e3-4f54-90c9-f8d714e1df40" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.211045] env[62385]: DEBUG nova.network.neutron [req-fb047e78-3cf4-4d01-bbe4-83f54fa18a7d req-6d19663f-094a-4a35-972f-0a2bcc764710 service nova] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Refreshing network info cache for port e2a123f6-199e-4e54-bd44-8b64435eebca {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 934.212517] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d3:20:43', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5116f690-f825-4fee-8a47-42b073e716c5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '89da5ce4-4773-4c11-91a4-870768a5b165', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:6d:a5:74', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a64108f9-df0a-4feb-bbb5-97f5841c356c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e2a123f6-199e-4e54-bd44-8b64435eebca', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 934.225562] env[62385]: DEBUG oslo.service.loopingcall [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 934.230063] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 934.230395] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-e75e6879-4fef-42a9-ba97-4fe18524ff8a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 934.230862] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-e75e6879-4fef-42a9-ba97-4fe18524ff8a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 934.231170] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-e75e6879-4fef-42a9-ba97-4fe18524ff8a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Deleting the datastore file [datastore1] caab700b-bd3c-497e-b40a-eb9d20dfc1e2 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 934.231871] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dfb19ccf-a696-48e1-9baa-38a3bd45e2fe {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.249572] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-725a9407-c300-44d6-acb1-8743e72d32e1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.257346] env[62385]: DEBUG oslo_vmware.api [None req-e75e6879-4fef-42a9-ba97-4fe18524ff8a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 934.257346] env[62385]: value = "task-1206136" [ 934.257346] env[62385]: _type = "Task" [ 934.257346] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.258553] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 934.258553] env[62385]: value = "task-1206137" [ 934.258553] env[62385]: _type = "Task" [ 934.258553] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.270909] env[62385]: DEBUG oslo_vmware.api [None req-e75e6879-4fef-42a9-ba97-4fe18524ff8a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206136, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.274044] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206137, 'name': CreateVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.368884] env[62385]: DEBUG nova.scheduler.client.report [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 934.493509] env[62385]: DEBUG oslo_vmware.api [None req-0d3ddba5-96d8-4d1f-a0cb-bb6183881104 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206134, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.472314} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.493509] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d3ddba5-96d8-4d1f-a0cb-bb6183881104 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 934.493509] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0d3ddba5-96d8-4d1f-a0cb-bb6183881104 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 934.493509] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0d3ddba5-96d8-4d1f-a0cb-bb6183881104 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 934.493669] env[62385]: INFO nova.compute.manager [None req-0d3ddba5-96d8-4d1f-a0cb-bb6183881104 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Took 0.60 seconds to destroy the instance on the hypervisor. [ 934.493910] env[62385]: DEBUG oslo.service.loopingcall [None req-0d3ddba5-96d8-4d1f-a0cb-bb6183881104 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 934.494730] env[62385]: DEBUG nova.compute.manager [-] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 934.495730] env[62385]: DEBUG nova.network.neutron [-] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 934.629678] env[62385]: DEBUG nova.network.neutron [req-fb047e78-3cf4-4d01-bbe4-83f54fa18a7d req-6d19663f-094a-4a35-972f-0a2bcc764710 service nova] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Updated VIF entry in instance network info cache for port e2a123f6-199e-4e54-bd44-8b64435eebca. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 934.630173] env[62385]: DEBUG nova.network.neutron [req-fb047e78-3cf4-4d01-bbe4-83f54fa18a7d req-6d19663f-094a-4a35-972f-0a2bcc764710 service nova] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Updating instance_info_cache with network_info: [{"id": "89da5ce4-4773-4c11-91a4-870768a5b165", "address": "fa:16:3e:d3:20:43", "network": {"id": "58340331-dbdb-4903-bab1-f17d05e248d3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1767092846", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.56", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d68661a0197d4e3cacc06ea433ac706e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5116f690-f825-4fee-8a47-42b073e716c5", "external-id": "nsx-vlan-transportzone-692", "segmentation_id": 692, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap89da5ce4-47", "ovs_interfaceid": "89da5ce4-4773-4c11-91a4-870768a5b165", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e2a123f6-199e-4e54-bd44-8b64435eebca", "address": "fa:16:3e:6d:a5:74", "network": {"id": "bba82320-3049-4f4a-a76e-52acce8b75ee", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-310743232", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.149", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "d68661a0197d4e3cacc06ea433ac706e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a64108f9-df0a-4feb-bbb5-97f5841c356c", "external-id": "nsx-vlan-transportzone-67", "segmentation_id": 67, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape2a123f6-19", "ovs_interfaceid": "e2a123f6-199e-4e54-bd44-8b64435eebca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.773335] env[62385]: DEBUG oslo_vmware.api [None req-e75e6879-4fef-42a9-ba97-4fe18524ff8a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206136, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.778898] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206137, 'name': CreateVM_Task, 'duration_secs': 0.433535} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.782717] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 934.782717] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.782717] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.782717] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 934.782717] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1921007b-625c-4f2d-89f2-94ad7576dae4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.785517] env[62385]: DEBUG oslo_vmware.api [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Waiting for the task: (returnval){ [ 934.785517] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]522f5423-7639-10c2-5d43-bd5b770a8bf2" [ 934.785517] env[62385]: _type = "Task" [ 934.785517] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.793667] env[62385]: DEBUG oslo_vmware.api [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]522f5423-7639-10c2-5d43-bd5b770a8bf2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.873848] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.411s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.874463] env[62385]: DEBUG nova.compute.manager [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 934.877182] env[62385]: DEBUG oslo_concurrency.lockutils [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.354s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.878552] env[62385]: INFO nova.compute.claims [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 935.134476] env[62385]: DEBUG oslo_concurrency.lockutils [req-fb047e78-3cf4-4d01-bbe4-83f54fa18a7d req-6d19663f-094a-4a35-972f-0a2bcc764710 service nova] Releasing lock "refresh_cache-75cf7907-d9e3-4f54-90c9-f8d714e1df40" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.270702] env[62385]: DEBUG oslo_vmware.api [None req-e75e6879-4fef-42a9-ba97-4fe18524ff8a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206136, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.684548} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.271071] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-e75e6879-4fef-42a9-ba97-4fe18524ff8a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 935.271186] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-e75e6879-4fef-42a9-ba97-4fe18524ff8a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 935.271368] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-e75e6879-4fef-42a9-ba97-4fe18524ff8a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 935.271549] env[62385]: INFO nova.compute.manager [None req-e75e6879-4fef-42a9-ba97-4fe18524ff8a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Took 1.72 seconds to destroy the instance on the hypervisor. [ 935.271793] env[62385]: DEBUG oslo.service.loopingcall [None req-e75e6879-4fef-42a9-ba97-4fe18524ff8a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 935.272013] env[62385]: DEBUG nova.compute.manager [-] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 935.272124] env[62385]: DEBUG nova.network.neutron [-] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 935.297095] env[62385]: DEBUG oslo_vmware.api [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]522f5423-7639-10c2-5d43-bd5b770a8bf2, 'name': SearchDatastore_Task, 'duration_secs': 0.0092} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.297538] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.297841] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 935.298103] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.298263] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.298458] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 935.298756] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-33bf56a2-3694-4e0b-be48-beab8d7cf750 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.307029] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 935.307178] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 935.308375] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e7e183f-ec2c-4a26-a982-c845a608d989 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.315779] env[62385]: DEBUG oslo_vmware.api [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Waiting for the task: (returnval){ [ 935.315779] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]526e9c4d-a3a7-4bc8-b21f-fe24fd005b83" [ 935.315779] env[62385]: _type = "Task" [ 935.315779] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.322840] env[62385]: DEBUG oslo_vmware.api [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]526e9c4d-a3a7-4bc8-b21f-fe24fd005b83, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.384048] env[62385]: DEBUG nova.compute.utils [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 935.384214] env[62385]: DEBUG nova.compute.manager [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 935.384384] env[62385]: DEBUG nova.network.neutron [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 935.479988] env[62385]: DEBUG nova.policy [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e56d4575033b4338acd01cdfa29f9693', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '09c2c5e938bf46fd9353e47aecc72478', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 935.636834] env[62385]: DEBUG nova.network.neutron [-] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.827442] env[62385]: DEBUG oslo_vmware.api [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]526e9c4d-a3a7-4bc8-b21f-fe24fd005b83, 'name': SearchDatastore_Task, 'duration_secs': 0.010671} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.828738] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d8c9b90-0714-42b3-a867-1f530d4cfc95 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.834487] env[62385]: DEBUG oslo_vmware.api [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Waiting for the task: (returnval){ [ 935.834487] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ef1165-094a-1ee8-2c1d-b46ef3fac9be" [ 935.834487] env[62385]: _type = "Task" [ 935.834487] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.842355] env[62385]: DEBUG oslo_vmware.api [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ef1165-094a-1ee8-2c1d-b46ef3fac9be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.895008] env[62385]: DEBUG nova.compute.manager [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 935.897517] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Acquiring lock "104dea2c-6977-4683-bc32-05131858f562" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.897746] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Lock "104dea2c-6977-4683-bc32-05131858f562" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.951430] env[62385]: DEBUG nova.network.neutron [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Successfully created port: deafdeb9-f59c-48f4-873b-6e4b27d4a688 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 936.063308] env[62385]: DEBUG nova.compute.manager [req-56b30998-bb4c-4cd7-8e88-aa0d772df17f req-8c375d0e-a42b-4e12-9b60-4a8f1515eed7 service nova] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Received event network-vif-deleted-f09e5aed-7121-4aae-b8e1-e6d84339486f {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 936.063308] env[62385]: DEBUG nova.compute.manager [req-56b30998-bb4c-4cd7-8e88-aa0d772df17f req-8c375d0e-a42b-4e12-9b60-4a8f1515eed7 service nova] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Received event network-vif-deleted-43db078b-0071-44a3-ab42-476acd4c56bf {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 936.063308] env[62385]: INFO nova.compute.manager [req-56b30998-bb4c-4cd7-8e88-aa0d772df17f req-8c375d0e-a42b-4e12-9b60-4a8f1515eed7 service nova] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Neutron deleted interface 43db078b-0071-44a3-ab42-476acd4c56bf; detaching it from the instance and deleting it from the info cache [ 936.063507] env[62385]: DEBUG nova.network.neutron [req-56b30998-bb4c-4cd7-8e88-aa0d772df17f req-8c375d0e-a42b-4e12-9b60-4a8f1515eed7 service nova] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.139081] env[62385]: INFO nova.compute.manager [-] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Took 1.64 seconds to deallocate network for instance. [ 936.198777] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6036099a-ee3b-40ca-b4a4-6f3373d41247 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.206822] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f92bb8f1-b1a3-4cb1-b121-b2fc9fdad91b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.238459] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f13144d6-a6ca-412f-8fde-016ea9431743 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.246304] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e6c45a5-7c46-446d-bab5-ca170ff845cd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.259069] env[62385]: DEBUG nova.network.neutron [-] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.260469] env[62385]: DEBUG nova.compute.provider_tree [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 936.344704] env[62385]: DEBUG oslo_vmware.api [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ef1165-094a-1ee8-2c1d-b46ef3fac9be, 'name': SearchDatastore_Task, 'duration_secs': 0.009056} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.344990] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.345268] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 75cf7907-d9e3-4f54-90c9-f8d714e1df40/75cf7907-d9e3-4f54-90c9-f8d714e1df40.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 936.345533] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-432f0fe6-cc77-4a47-93ae-4fb6d21e2893 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.353217] env[62385]: DEBUG oslo_vmware.api [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Waiting for the task: (returnval){ [ 936.353217] env[62385]: value = "task-1206138" [ 936.353217] env[62385]: _type = "Task" [ 936.353217] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.361113] env[62385]: DEBUG oslo_vmware.api [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': task-1206138, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.403932] env[62385]: DEBUG nova.compute.manager [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] [instance: 104dea2c-6977-4683-bc32-05131858f562] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 936.567763] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8dc09cb0-8860-4dbc-9a02-a12e9ce0b6a7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.581923] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1afbc0fa-bc29-4c8d-88d5-82966aa600a7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.622687] env[62385]: DEBUG nova.compute.manager [req-56b30998-bb4c-4cd7-8e88-aa0d772df17f req-8c375d0e-a42b-4e12-9b60-4a8f1515eed7 service nova] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Detach interface failed, port_id=43db078b-0071-44a3-ab42-476acd4c56bf, reason: Instance caab700b-bd3c-497e-b40a-eb9d20dfc1e2 could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 936.647105] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0d3ddba5-96d8-4d1f-a0cb-bb6183881104 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.762705] env[62385]: INFO nova.compute.manager [-] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Took 1.49 seconds to deallocate network for instance. [ 936.763621] env[62385]: DEBUG nova.scheduler.client.report [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 936.863048] env[62385]: DEBUG oslo_vmware.api [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': task-1206138, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.482438} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.863361] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 75cf7907-d9e3-4f54-90c9-f8d714e1df40/75cf7907-d9e3-4f54-90c9-f8d714e1df40.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 936.864384] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 936.864384] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d0414f0c-ecfd-4d37-b086-cb38e2044208 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.871147] env[62385]: DEBUG oslo_vmware.api [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Waiting for the task: (returnval){ [ 936.871147] env[62385]: value = "task-1206139" [ 936.871147] env[62385]: _type = "Task" [ 936.871147] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.879041] env[62385]: DEBUG oslo_vmware.api [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': task-1206139, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.907855] env[62385]: DEBUG nova.compute.manager [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 936.929664] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.941253] env[62385]: DEBUG nova.virt.hardware [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 936.941515] env[62385]: DEBUG nova.virt.hardware [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 936.941687] env[62385]: DEBUG nova.virt.hardware [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 936.941869] env[62385]: DEBUG nova.virt.hardware [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 936.942090] env[62385]: DEBUG nova.virt.hardware [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 936.942271] env[62385]: DEBUG nova.virt.hardware [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 936.942484] env[62385]: DEBUG nova.virt.hardware [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 936.942648] env[62385]: DEBUG nova.virt.hardware [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 936.942820] env[62385]: DEBUG nova.virt.hardware [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 936.942986] env[62385]: DEBUG nova.virt.hardware [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 936.943175] env[62385]: DEBUG nova.virt.hardware [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 936.944034] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad26d75a-2db7-43aa-a10c-dee801f92a95 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.952498] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-610cf364-862c-4382-a560-fe8c21f67ed4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.229329] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fea8ac26-d2cb-41ad-8f71-07e51195c9f1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Acquiring lock "d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.229703] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fea8ac26-d2cb-41ad-8f71-07e51195c9f1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.230753] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fea8ac26-d2cb-41ad-8f71-07e51195c9f1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Acquiring lock "d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.230957] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fea8ac26-d2cb-41ad-8f71-07e51195c9f1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.231174] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fea8ac26-d2cb-41ad-8f71-07e51195c9f1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.234602] env[62385]: INFO nova.compute.manager [None req-fea8ac26-d2cb-41ad-8f71-07e51195c9f1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Terminating instance [ 937.236431] env[62385]: DEBUG nova.compute.manager [None req-fea8ac26-d2cb-41ad-8f71-07e51195c9f1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 937.236631] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fea8ac26-d2cb-41ad-8f71-07e51195c9f1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 937.237466] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fbafc79-419b-4533-84a9-2428729c35d8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.245169] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fea8ac26-d2cb-41ad-8f71-07e51195c9f1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 937.245406] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f696e4a4-ef75-4fc2-a750-f46b00404703 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.251915] env[62385]: DEBUG oslo_vmware.api [None req-fea8ac26-d2cb-41ad-8f71-07e51195c9f1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for the task: (returnval){ [ 937.251915] env[62385]: value = "task-1206140" [ 937.251915] env[62385]: _type = "Task" [ 937.251915] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.260164] env[62385]: DEBUG oslo_vmware.api [None req-fea8ac26-d2cb-41ad-8f71-07e51195c9f1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206140, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.272248] env[62385]: DEBUG oslo_concurrency.lockutils [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.395s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.272799] env[62385]: DEBUG nova.compute.manager [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 937.275658] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c0464272-cede-4d75-8256-7a60bb9810a1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.561s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.275992] env[62385]: DEBUG nova.objects.instance [None req-c0464272-cede-4d75-8256-7a60bb9810a1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Lazy-loading 'resources' on Instance uuid 09c017a9-ad18-49b6-b72d-958023c81b24 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 937.278203] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e75e6879-4fef-42a9-ba97-4fe18524ff8a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.381784] env[62385]: DEBUG oslo_vmware.api [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': task-1206139, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066952} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.382118] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 937.383233] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-606501ce-a64d-48dc-b635-e23565c1a6ac {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.409771] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] 75cf7907-d9e3-4f54-90c9-f8d714e1df40/75cf7907-d9e3-4f54-90c9-f8d714e1df40.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 937.410096] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8e88cab5-f2d7-4e20-abc1-0ab46c977d3f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.430808] env[62385]: DEBUG oslo_vmware.api [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Waiting for the task: (returnval){ [ 937.430808] env[62385]: value = "task-1206141" [ 937.430808] env[62385]: _type = "Task" [ 937.430808] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.441524] env[62385]: DEBUG oslo_vmware.api [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': task-1206141, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.764155] env[62385]: DEBUG nova.network.neutron [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Successfully updated port: deafdeb9-f59c-48f4-873b-6e4b27d4a688 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 937.773545] env[62385]: DEBUG oslo_vmware.api [None req-fea8ac26-d2cb-41ad-8f71-07e51195c9f1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206140, 'name': PowerOffVM_Task, 'duration_secs': 0.206851} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.774362] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-fea8ac26-d2cb-41ad-8f71-07e51195c9f1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 937.774362] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fea8ac26-d2cb-41ad-8f71-07e51195c9f1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 937.774362] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-44ef2ded-bbba-4265-a0cc-29e579ee975a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.781372] env[62385]: DEBUG nova.compute.utils [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 937.782805] env[62385]: DEBUG nova.compute.manager [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 937.783220] env[62385]: DEBUG nova.network.neutron [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 937.823203] env[62385]: DEBUG nova.policy [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5e83312d226945c99d05fcc7f74c0978', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bf8e3c942d3445919cfbe988cca84e90', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 937.842782] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fea8ac26-d2cb-41ad-8f71-07e51195c9f1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 937.843045] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fea8ac26-d2cb-41ad-8f71-07e51195c9f1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 937.843205] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-fea8ac26-d2cb-41ad-8f71-07e51195c9f1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Deleting the datastore file [datastore1] d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 937.843469] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1d6f3005-a1d5-4b5b-9850-4837e83e2c3d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.855871] env[62385]: DEBUG oslo_vmware.api [None req-fea8ac26-d2cb-41ad-8f71-07e51195c9f1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for the task: (returnval){ [ 937.855871] env[62385]: value = "task-1206143" [ 937.855871] env[62385]: _type = "Task" [ 937.855871] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.864713] env[62385]: DEBUG oslo_vmware.api [None req-fea8ac26-d2cb-41ad-8f71-07e51195c9f1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206143, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.942937] env[62385]: DEBUG oslo_vmware.api [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': task-1206141, 'name': ReconfigVM_Task, 'duration_secs': 0.278057} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.943148] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Reconfigured VM instance instance-0000005a to attach disk [datastore2] 75cf7907-d9e3-4f54-90c9-f8d714e1df40/75cf7907-d9e3-4f54-90c9-f8d714e1df40.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 937.944462] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f8571c27-e127-4840-a899-1c473f857aeb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.952345] env[62385]: DEBUG oslo_vmware.api [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Waiting for the task: (returnval){ [ 937.952345] env[62385]: value = "task-1206144" [ 937.952345] env[62385]: _type = "Task" [ 937.952345] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.960045] env[62385]: DEBUG oslo_vmware.api [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': task-1206144, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.075196] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a12f1afa-9b1b-4c5b-a2bb-a294058e18cb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.082405] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-741ee669-9c72-4a21-a784-85806686c955 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.113199] env[62385]: DEBUG nova.compute.manager [req-e8509eeb-6441-479e-9688-60379d657000 req-f912ab8b-a67a-4c03-bb0e-c4ca733788ea service nova] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Received event network-vif-plugged-deafdeb9-f59c-48f4-873b-6e4b27d4a688 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 938.113433] env[62385]: DEBUG oslo_concurrency.lockutils [req-e8509eeb-6441-479e-9688-60379d657000 req-f912ab8b-a67a-4c03-bb0e-c4ca733788ea service nova] Acquiring lock "8e1d416c-2624-468c-94d7-6265e9f4178f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.113990] env[62385]: DEBUG oslo_concurrency.lockutils [req-e8509eeb-6441-479e-9688-60379d657000 req-f912ab8b-a67a-4c03-bb0e-c4ca733788ea service nova] Lock "8e1d416c-2624-468c-94d7-6265e9f4178f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.114268] env[62385]: DEBUG oslo_concurrency.lockutils [req-e8509eeb-6441-479e-9688-60379d657000 req-f912ab8b-a67a-4c03-bb0e-c4ca733788ea service nova] Lock "8e1d416c-2624-468c-94d7-6265e9f4178f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.114473] env[62385]: DEBUG nova.compute.manager [req-e8509eeb-6441-479e-9688-60379d657000 req-f912ab8b-a67a-4c03-bb0e-c4ca733788ea service nova] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] No waiting events found dispatching network-vif-plugged-deafdeb9-f59c-48f4-873b-6e4b27d4a688 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 938.114778] env[62385]: WARNING nova.compute.manager [req-e8509eeb-6441-479e-9688-60379d657000 req-f912ab8b-a67a-4c03-bb0e-c4ca733788ea service nova] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Received unexpected event network-vif-plugged-deafdeb9-f59c-48f4-873b-6e4b27d4a688 for instance with vm_state building and task_state spawning. [ 938.114961] env[62385]: DEBUG nova.compute.manager [req-e8509eeb-6441-479e-9688-60379d657000 req-f912ab8b-a67a-4c03-bb0e-c4ca733788ea service nova] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Received event network-changed-deafdeb9-f59c-48f4-873b-6e4b27d4a688 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 938.115140] env[62385]: DEBUG nova.compute.manager [req-e8509eeb-6441-479e-9688-60379d657000 req-f912ab8b-a67a-4c03-bb0e-c4ca733788ea service nova] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Refreshing instance network info cache due to event network-changed-deafdeb9-f59c-48f4-873b-6e4b27d4a688. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 938.115340] env[62385]: DEBUG oslo_concurrency.lockutils [req-e8509eeb-6441-479e-9688-60379d657000 req-f912ab8b-a67a-4c03-bb0e-c4ca733788ea service nova] Acquiring lock "refresh_cache-8e1d416c-2624-468c-94d7-6265e9f4178f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.115481] env[62385]: DEBUG oslo_concurrency.lockutils [req-e8509eeb-6441-479e-9688-60379d657000 req-f912ab8b-a67a-4c03-bb0e-c4ca733788ea service nova] Acquired lock "refresh_cache-8e1d416c-2624-468c-94d7-6265e9f4178f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.115640] env[62385]: DEBUG nova.network.neutron [req-e8509eeb-6441-479e-9688-60379d657000 req-f912ab8b-a67a-4c03-bb0e-c4ca733788ea service nova] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Refreshing network info cache for port deafdeb9-f59c-48f4-873b-6e4b27d4a688 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 938.119727] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c2af9ec-a9be-4602-832d-2463189519fe {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.126845] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e42c2ee-cc16-4d2f-9a83-b37c2be306df {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.147675] env[62385]: DEBUG nova.compute.provider_tree [None req-c0464272-cede-4d75-8256-7a60bb9810a1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 938.151213] env[62385]: DEBUG nova.network.neutron [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Successfully created port: 0533b6c1-8d5b-4822-acd6-a95a651ba50c {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 938.266394] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquiring lock "refresh_cache-8e1d416c-2624-468c-94d7-6265e9f4178f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.286432] env[62385]: DEBUG nova.compute.manager [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 938.366518] env[62385]: DEBUG oslo_vmware.api [None req-fea8ac26-d2cb-41ad-8f71-07e51195c9f1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206143, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13801} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.366774] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-fea8ac26-d2cb-41ad-8f71-07e51195c9f1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 938.367014] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fea8ac26-d2cb-41ad-8f71-07e51195c9f1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 938.367347] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fea8ac26-d2cb-41ad-8f71-07e51195c9f1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 938.367564] env[62385]: INFO nova.compute.manager [None req-fea8ac26-d2cb-41ad-8f71-07e51195c9f1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Took 1.13 seconds to destroy the instance on the hypervisor. [ 938.367818] env[62385]: DEBUG oslo.service.loopingcall [None req-fea8ac26-d2cb-41ad-8f71-07e51195c9f1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 938.368037] env[62385]: DEBUG nova.compute.manager [-] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 938.368137] env[62385]: DEBUG nova.network.neutron [-] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 938.463014] env[62385]: DEBUG oslo_vmware.api [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': task-1206144, 'name': Rename_Task, 'duration_secs': 0.140018} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.463426] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 938.463682] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c20b4985-a57c-4612-a9d6-0e34b869616e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.471855] env[62385]: DEBUG oslo_vmware.api [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Waiting for the task: (returnval){ [ 938.471855] env[62385]: value = "task-1206145" [ 938.471855] env[62385]: _type = "Task" [ 938.471855] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.479971] env[62385]: DEBUG oslo_vmware.api [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': task-1206145, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.651115] env[62385]: DEBUG nova.scheduler.client.report [None req-c0464272-cede-4d75-8256-7a60bb9810a1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 938.658930] env[62385]: DEBUG nova.network.neutron [req-e8509eeb-6441-479e-9688-60379d657000 req-f912ab8b-a67a-4c03-bb0e-c4ca733788ea service nova] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 938.744934] env[62385]: DEBUG nova.network.neutron [req-e8509eeb-6441-479e-9688-60379d657000 req-f912ab8b-a67a-4c03-bb0e-c4ca733788ea service nova] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.989069] env[62385]: DEBUG oslo_vmware.api [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': task-1206145, 'name': PowerOnVM_Task, 'duration_secs': 0.47118} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.989346] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 938.989612] env[62385]: INFO nova.compute.manager [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Took 11.10 seconds to spawn the instance on the hypervisor. [ 938.989808] env[62385]: DEBUG nova.compute.manager [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 938.990752] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aae798b6-d323-44e0-a1e9-1b3cdd3330af {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.163022] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c0464272-cede-4d75-8256-7a60bb9810a1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.887s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.166182] env[62385]: DEBUG oslo_concurrency.lockutils [None req-53444455-1881-4ba7-b70c-4c868d7f0a6e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 17.370s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.188772] env[62385]: INFO nova.scheduler.client.report [None req-c0464272-cede-4d75-8256-7a60bb9810a1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Deleted allocations for instance 09c017a9-ad18-49b6-b72d-958023c81b24 [ 939.247691] env[62385]: DEBUG oslo_concurrency.lockutils [req-e8509eeb-6441-479e-9688-60379d657000 req-f912ab8b-a67a-4c03-bb0e-c4ca733788ea service nova] Releasing lock "refresh_cache-8e1d416c-2624-468c-94d7-6265e9f4178f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.250185] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquired lock "refresh_cache-8e1d416c-2624-468c-94d7-6265e9f4178f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.250185] env[62385]: DEBUG nova.network.neutron [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 939.299214] env[62385]: DEBUG nova.compute.manager [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 939.330108] env[62385]: DEBUG nova.virt.hardware [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 939.330108] env[62385]: DEBUG nova.virt.hardware [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 939.330108] env[62385]: DEBUG nova.virt.hardware [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 939.330108] env[62385]: DEBUG nova.virt.hardware [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 939.330108] env[62385]: DEBUG nova.virt.hardware [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 939.330108] env[62385]: DEBUG nova.virt.hardware [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 939.330475] env[62385]: DEBUG nova.virt.hardware [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 939.330657] env[62385]: DEBUG nova.virt.hardware [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 939.330728] env[62385]: DEBUG nova.virt.hardware [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 939.330893] env[62385]: DEBUG nova.virt.hardware [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 939.331084] env[62385]: DEBUG nova.virt.hardware [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 939.331976] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef362b81-4f4e-4c63-b7e0-be8dc6ede717 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.340678] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3332f2ca-b8ad-4ca2-9d23-9c8ee5930c2a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.509212] env[62385]: INFO nova.compute.manager [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Took 38.59 seconds to build instance. [ 939.661259] env[62385]: DEBUG nova.compute.manager [req-6e55fa49-02bf-4c28-bbe9-ff0b2daf2b20 req-b4858f20-4a3a-4dff-ae1e-11f69100eaac service nova] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Received event network-vif-plugged-0533b6c1-8d5b-4822-acd6-a95a651ba50c {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 939.661792] env[62385]: DEBUG oslo_concurrency.lockutils [req-6e55fa49-02bf-4c28-bbe9-ff0b2daf2b20 req-b4858f20-4a3a-4dff-ae1e-11f69100eaac service nova] Acquiring lock "df559cb8-555c-496b-bce3-0981be3ef65f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.661792] env[62385]: DEBUG oslo_concurrency.lockutils [req-6e55fa49-02bf-4c28-bbe9-ff0b2daf2b20 req-b4858f20-4a3a-4dff-ae1e-11f69100eaac service nova] Lock "df559cb8-555c-496b-bce3-0981be3ef65f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.661895] env[62385]: DEBUG oslo_concurrency.lockutils [req-6e55fa49-02bf-4c28-bbe9-ff0b2daf2b20 req-b4858f20-4a3a-4dff-ae1e-11f69100eaac service nova] Lock "df559cb8-555c-496b-bce3-0981be3ef65f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.664264] env[62385]: DEBUG nova.compute.manager [req-6e55fa49-02bf-4c28-bbe9-ff0b2daf2b20 req-b4858f20-4a3a-4dff-ae1e-11f69100eaac service nova] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] No waiting events found dispatching network-vif-plugged-0533b6c1-8d5b-4822-acd6-a95a651ba50c {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 939.664264] env[62385]: WARNING nova.compute.manager [req-6e55fa49-02bf-4c28-bbe9-ff0b2daf2b20 req-b4858f20-4a3a-4dff-ae1e-11f69100eaac service nova] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Received unexpected event network-vif-plugged-0533b6c1-8d5b-4822-acd6-a95a651ba50c for instance with vm_state building and task_state spawning. [ 939.706619] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c0464272-cede-4d75-8256-7a60bb9810a1 tempest-ImagesOneServerNegativeTestJSON-1833813522 tempest-ImagesOneServerNegativeTestJSON-1833813522-project-member] Lock "09c017a9-ad18-49b6-b72d-958023c81b24" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.886s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.768384] env[62385]: DEBUG nova.network.neutron [-] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.816194] env[62385]: DEBUG nova.network.neutron [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 939.962607] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8dddf89-a1d2-4014-b8e1-73a52c24afcc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.970916] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ffd9a04-92d7-4fc1-8bc7-dff50b5d9ec6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.003873] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2ebad86-ed5a-4dec-a6c2-be2ffd639963 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.011215] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5966d371-e779-42d8-ba6b-514b110efc83 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Lock "75cf7907-d9e3-4f54-90c9-f8d714e1df40" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.100s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.012181] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaca6648-c969-472d-8804-1d1f78c5ec5f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.028702] env[62385]: DEBUG nova.compute.provider_tree [None req-53444455-1881-4ba7-b70c-4c868d7f0a6e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 940.041973] env[62385]: DEBUG nova.network.neutron [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Updating instance_info_cache with network_info: [{"id": "deafdeb9-f59c-48f4-873b-6e4b27d4a688", "address": "fa:16:3e:81:4d:75", "network": {"id": "9c4d6464-cacc-44f8-b7e5-0699cd312c53", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1889153615-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09c2c5e938bf46fd9353e47aecc72478", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4d3f69a-b086-4c3b-b976-5a848b63dfc4", "external-id": "nsx-vlan-transportzone-627", "segmentation_id": 627, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdeafdeb9-f5", "ovs_interfaceid": "deafdeb9-f59c-48f4-873b-6e4b27d4a688", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.193272] env[62385]: DEBUG nova.compute.manager [req-070f01be-f046-499a-bb7b-b7329368d059 req-992a6ee0-35cf-45df-a63d-88424b9b7916 service nova] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Received event network-vif-deleted-2c31d161-814f-4c23-9882-0aa436bb9e03 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 940.258069] env[62385]: DEBUG nova.network.neutron [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Successfully updated port: 0533b6c1-8d5b-4822-acd6-a95a651ba50c {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 940.272824] env[62385]: INFO nova.compute.manager [-] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Took 1.90 seconds to deallocate network for instance. [ 940.532138] env[62385]: DEBUG nova.scheduler.client.report [None req-53444455-1881-4ba7-b70c-4c868d7f0a6e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 940.544346] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Releasing lock "refresh_cache-8e1d416c-2624-468c-94d7-6265e9f4178f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.544669] env[62385]: DEBUG nova.compute.manager [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Instance network_info: |[{"id": "deafdeb9-f59c-48f4-873b-6e4b27d4a688", "address": "fa:16:3e:81:4d:75", "network": {"id": "9c4d6464-cacc-44f8-b7e5-0699cd312c53", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1889153615-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09c2c5e938bf46fd9353e47aecc72478", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4d3f69a-b086-4c3b-b976-5a848b63dfc4", "external-id": "nsx-vlan-transportzone-627", "segmentation_id": 627, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdeafdeb9-f5", "ovs_interfaceid": "deafdeb9-f59c-48f4-873b-6e4b27d4a688", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 940.545120] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:81:4d:75', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c4d3f69a-b086-4c3b-b976-5a848b63dfc4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'deafdeb9-f59c-48f4-873b-6e4b27d4a688', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 940.553397] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Creating folder: Project (09c2c5e938bf46fd9353e47aecc72478). Parent ref: group-v261107. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 940.554403] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-73cdfec4-2a09-489d-9f97-4ab45aa9e20e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.565615] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Created folder: Project (09c2c5e938bf46fd9353e47aecc72478) in parent group-v261107. [ 940.565815] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Creating folder: Instances. Parent ref: group-v261260. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 940.566068] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6c142688-4f5c-4e54-9432-9c6b4a0f014a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.574236] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Created folder: Instances in parent group-v261260. [ 940.574468] env[62385]: DEBUG oslo.service.loopingcall [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 940.574682] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 940.574853] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fe76442a-ca9d-428a-be84-b407ecf47eb2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.593451] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 940.593451] env[62385]: value = "task-1206148" [ 940.593451] env[62385]: _type = "Task" [ 940.593451] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.604400] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206148, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.717457] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "fd0b59c2-e38d-4ecc-b32f-72f07d555ffd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.717712] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "fd0b59c2-e38d-4ecc-b32f-72f07d555ffd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.759605] env[62385]: DEBUG oslo_concurrency.lockutils [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "refresh_cache-df559cb8-555c-496b-bce3-0981be3ef65f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.759764] env[62385]: DEBUG oslo_concurrency.lockutils [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquired lock "refresh_cache-df559cb8-555c-496b-bce3-0981be3ef65f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.759918] env[62385]: DEBUG nova.network.neutron [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 940.779055] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fea8ac26-d2cb-41ad-8f71-07e51195c9f1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.802763] env[62385]: DEBUG oslo_concurrency.lockutils [None req-84b85202-2a88-4ae6-a721-84e76137e9f4 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Acquiring lock "75cf7907-d9e3-4f54-90c9-f8d714e1df40" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.802982] env[62385]: DEBUG oslo_concurrency.lockutils [None req-84b85202-2a88-4ae6-a721-84e76137e9f4 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Lock "75cf7907-d9e3-4f54-90c9-f8d714e1df40" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.803182] env[62385]: DEBUG oslo_concurrency.lockutils [None req-84b85202-2a88-4ae6-a721-84e76137e9f4 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Acquiring lock "75cf7907-d9e3-4f54-90c9-f8d714e1df40-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.803466] env[62385]: DEBUG oslo_concurrency.lockutils [None req-84b85202-2a88-4ae6-a721-84e76137e9f4 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Lock "75cf7907-d9e3-4f54-90c9-f8d714e1df40-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.803585] env[62385]: DEBUG oslo_concurrency.lockutils [None req-84b85202-2a88-4ae6-a721-84e76137e9f4 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Lock "75cf7907-d9e3-4f54-90c9-f8d714e1df40-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.807832] env[62385]: INFO nova.compute.manager [None req-84b85202-2a88-4ae6-a721-84e76137e9f4 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Terminating instance [ 940.809914] env[62385]: DEBUG nova.compute.manager [None req-84b85202-2a88-4ae6-a721-84e76137e9f4 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 940.810135] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-84b85202-2a88-4ae6-a721-84e76137e9f4 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 940.811063] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b341edc9-c19f-472a-816e-52d1fc9f1a9f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.820187] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-84b85202-2a88-4ae6-a721-84e76137e9f4 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 940.820443] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ccbf41f5-1ebb-4908-9fda-aa88ad5e124f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.826867] env[62385]: DEBUG oslo_vmware.api [None req-84b85202-2a88-4ae6-a721-84e76137e9f4 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Waiting for the task: (returnval){ [ 940.826867] env[62385]: value = "task-1206149" [ 940.826867] env[62385]: _type = "Task" [ 940.826867] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.835093] env[62385]: DEBUG oslo_vmware.api [None req-84b85202-2a88-4ae6-a721-84e76137e9f4 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': task-1206149, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.103818] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206148, 'name': CreateVM_Task, 'duration_secs': 0.345178} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.103943] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 941.105788] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.105967] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.106317] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 941.106574] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74e71c8b-abc2-43b8-bf8e-6dd0632bb24d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.111202] env[62385]: DEBUG oslo_vmware.api [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 941.111202] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c88af1-0773-0754-85d0-d8352fd52f90" [ 941.111202] env[62385]: _type = "Task" [ 941.111202] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.120485] env[62385]: DEBUG oslo_vmware.api [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c88af1-0773-0754-85d0-d8352fd52f90, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.220624] env[62385]: DEBUG nova.compute.manager [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 941.313911] env[62385]: DEBUG nova.network.neutron [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 941.336736] env[62385]: DEBUG oslo_vmware.api [None req-84b85202-2a88-4ae6-a721-84e76137e9f4 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': task-1206149, 'name': PowerOffVM_Task, 'duration_secs': 0.21183} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.341129] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-84b85202-2a88-4ae6-a721-84e76137e9f4 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 941.341322] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-84b85202-2a88-4ae6-a721-84e76137e9f4 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 941.341972] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d140c5ec-edc1-481a-af06-036ed43080db {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.445110] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-84b85202-2a88-4ae6-a721-84e76137e9f4 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 941.445110] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-84b85202-2a88-4ae6-a721-84e76137e9f4 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Deleting contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 941.445110] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-84b85202-2a88-4ae6-a721-84e76137e9f4 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Deleting the datastore file [datastore2] 75cf7907-d9e3-4f54-90c9-f8d714e1df40 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 941.445110] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-71169187-7d60-4ddb-9f98-2457e993e62b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.451529] env[62385]: DEBUG oslo_vmware.api [None req-84b85202-2a88-4ae6-a721-84e76137e9f4 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Waiting for the task: (returnval){ [ 941.451529] env[62385]: value = "task-1206151" [ 941.451529] env[62385]: _type = "Task" [ 941.451529] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.460393] env[62385]: DEBUG oslo_vmware.api [None req-84b85202-2a88-4ae6-a721-84e76137e9f4 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': task-1206151, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.542480] env[62385]: DEBUG oslo_concurrency.lockutils [None req-53444455-1881-4ba7-b70c-4c868d7f0a6e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.376s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.547915] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 18.081s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.627491] env[62385]: DEBUG oslo_vmware.api [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c88af1-0773-0754-85d0-d8352fd52f90, 'name': SearchDatastore_Task, 'duration_secs': 0.010138} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.627491] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.627491] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 941.627491] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.627491] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.627491] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 941.627491] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-09e18e48-1b3d-4c67-9c24-450e8b2768d5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.637030] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 941.637030] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 941.637030] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a130d609-e488-4c09-a8cd-8952e7eb10f3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.643865] env[62385]: DEBUG oslo_vmware.api [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 941.643865] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]520b9111-5f93-d05a-fe12-1fb8c30646ba" [ 941.643865] env[62385]: _type = "Task" [ 941.643865] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.648914] env[62385]: DEBUG oslo_vmware.api [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]520b9111-5f93-d05a-fe12-1fb8c30646ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.747680] env[62385]: DEBUG nova.network.neutron [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Updating instance_info_cache with network_info: [{"id": "0533b6c1-8d5b-4822-acd6-a95a651ba50c", "address": "fa:16:3e:db:e5:07", "network": {"id": "8b2d7a6e-25d4-4a54-ab54-5972612046da", "bridge": "br-int", "label": "tempest-ServersTestJSON-1142134211-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bf8e3c942d3445919cfbe988cca84e90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0533b6c1-8d", "ovs_interfaceid": "0533b6c1-8d5b-4822-acd6-a95a651ba50c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.754438] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.883205] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f1e1f61c-2527-4118-889d-0942f17b62b3 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "interface-f937ef7b-ab39-4d8a-9577-7faabc652aa5-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.883582] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f1e1f61c-2527-4118-889d-0942f17b62b3 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "interface-f937ef7b-ab39-4d8a-9577-7faabc652aa5-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.883923] env[62385]: DEBUG nova.objects.instance [None req-f1e1f61c-2527-4118-889d-0942f17b62b3 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lazy-loading 'flavor' on Instance uuid f937ef7b-ab39-4d8a-9577-7faabc652aa5 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 941.964640] env[62385]: DEBUG oslo_vmware.api [None req-84b85202-2a88-4ae6-a721-84e76137e9f4 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Task: {'id': task-1206151, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146923} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.964921] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-84b85202-2a88-4ae6-a721-84e76137e9f4 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 941.965124] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-84b85202-2a88-4ae6-a721-84e76137e9f4 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Deleted contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 941.965309] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-84b85202-2a88-4ae6-a721-84e76137e9f4 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 941.965485] env[62385]: INFO nova.compute.manager [None req-84b85202-2a88-4ae6-a721-84e76137e9f4 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Took 1.16 seconds to destroy the instance on the hypervisor. [ 941.965742] env[62385]: DEBUG oslo.service.loopingcall [None req-84b85202-2a88-4ae6-a721-84e76137e9f4 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 941.965920] env[62385]: DEBUG nova.compute.manager [-] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 941.966590] env[62385]: DEBUG nova.network.neutron [-] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 942.119496] env[62385]: INFO nova.scheduler.client.report [None req-53444455-1881-4ba7-b70c-4c868d7f0a6e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Deleted allocation for migration 5c8fd55d-d3a1-49b8-8f53-46b2748a8591 [ 942.154687] env[62385]: DEBUG oslo_vmware.api [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]520b9111-5f93-d05a-fe12-1fb8c30646ba, 'name': SearchDatastore_Task, 'duration_secs': 0.007734} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.155535] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd93843a-e7a8-4cc7-91c1-47004b83ffa4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.161249] env[62385]: DEBUG oslo_vmware.api [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 942.161249] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a6b005-5e59-e2a3-6088-2350e543243f" [ 942.161249] env[62385]: _type = "Task" [ 942.161249] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.170068] env[62385]: DEBUG oslo_vmware.api [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a6b005-5e59-e2a3-6088-2350e543243f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.251032] env[62385]: DEBUG oslo_concurrency.lockutils [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Releasing lock "refresh_cache-df559cb8-555c-496b-bce3-0981be3ef65f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.251349] env[62385]: DEBUG nova.compute.manager [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Instance network_info: |[{"id": "0533b6c1-8d5b-4822-acd6-a95a651ba50c", "address": "fa:16:3e:db:e5:07", "network": {"id": "8b2d7a6e-25d4-4a54-ab54-5972612046da", "bridge": "br-int", "label": "tempest-ServersTestJSON-1142134211-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bf8e3c942d3445919cfbe988cca84e90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0533b6c1-8d", "ovs_interfaceid": "0533b6c1-8d5b-4822-acd6-a95a651ba50c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 942.251805] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:db:e5:07', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0533b6c1-8d5b-4822-acd6-a95a651ba50c', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 942.261984] env[62385]: DEBUG oslo.service.loopingcall [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 942.262270] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 942.262499] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e4f78149-3bf6-46b4-b9da-7b45d1a5810a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.285616] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 942.285616] env[62385]: value = "task-1206152" [ 942.285616] env[62385]: _type = "Task" [ 942.285616] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.294759] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206152, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.388376] env[62385]: DEBUG nova.objects.instance [None req-f1e1f61c-2527-4118-889d-0942f17b62b3 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lazy-loading 'pci_requests' on Instance uuid f937ef7b-ab39-4d8a-9577-7faabc652aa5 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 942.588690] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 33387505-c576-488b-8c9c-b064fe81a7d7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 942.588690] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance f1d33401-35f5-4d79-abb4-26dc6faa784e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 942.588690] env[62385]: WARNING nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 942.588690] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 9d6f098a-0b05-43ef-96b0-9eb99ad3538c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 942.588690] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 942.588690] env[62385]: WARNING nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 679150f5-ad17-45e6-9e3d-9f6e28b1aee3 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 942.588690] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance bd7263b2-b996-4794-946b-2c28215574cb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 942.588690] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance f937ef7b-ab39-4d8a-9577-7faabc652aa5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 942.588690] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance a58cb15f-e4df-4b1e-b09c-2f61dce7200f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 942.588690] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 9b2f8292-9d89-407e-96c5-195ee398cc7b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 942.588690] env[62385]: WARNING nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance caab700b-bd3c-497e-b40a-eb9d20dfc1e2 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 942.588690] env[62385]: WARNING nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance fad3b054-cf2a-4ee3-a2a5-918a96a4a125 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 942.588690] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 75cf7907-d9e3-4f54-90c9-f8d714e1df40 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 942.589684] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 8e1d416c-2624-468c-94d7-6265e9f4178f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 942.589684] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance df559cb8-555c-496b-bce3-0981be3ef65f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 942.599808] env[62385]: DEBUG nova.compute.manager [req-91be97db-85c3-425f-86d0-97c8e5510af7 req-d32a7cf9-5bef-4e47-893b-51956e4c0907 service nova] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Received event network-changed-0533b6c1-8d5b-4822-acd6-a95a651ba50c {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 942.600010] env[62385]: DEBUG nova.compute.manager [req-91be97db-85c3-425f-86d0-97c8e5510af7 req-d32a7cf9-5bef-4e47-893b-51956e4c0907 service nova] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Refreshing instance network info cache due to event network-changed-0533b6c1-8d5b-4822-acd6-a95a651ba50c. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 942.600232] env[62385]: DEBUG oslo_concurrency.lockutils [req-91be97db-85c3-425f-86d0-97c8e5510af7 req-d32a7cf9-5bef-4e47-893b-51956e4c0907 service nova] Acquiring lock "refresh_cache-df559cb8-555c-496b-bce3-0981be3ef65f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.600381] env[62385]: DEBUG oslo_concurrency.lockutils [req-91be97db-85c3-425f-86d0-97c8e5510af7 req-d32a7cf9-5bef-4e47-893b-51956e4c0907 service nova] Acquired lock "refresh_cache-df559cb8-555c-496b-bce3-0981be3ef65f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.600616] env[62385]: DEBUG nova.network.neutron [req-91be97db-85c3-425f-86d0-97c8e5510af7 req-d32a7cf9-5bef-4e47-893b-51956e4c0907 service nova] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Refreshing network info cache for port 0533b6c1-8d5b-4822-acd6-a95a651ba50c {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 942.625635] env[62385]: DEBUG oslo_concurrency.lockutils [None req-53444455-1881-4ba7-b70c-4c868d7f0a6e tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "a58cb15f-e4df-4b1e-b09c-2f61dce7200f" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 23.698s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.671833] env[62385]: DEBUG oslo_vmware.api [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a6b005-5e59-e2a3-6088-2350e543243f, 'name': SearchDatastore_Task, 'duration_secs': 0.009916} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.672353] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.672620] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 8e1d416c-2624-468c-94d7-6265e9f4178f/8e1d416c-2624-468c-94d7-6265e9f4178f.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 942.672889] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-73d1ecaa-96e6-43de-8cdd-13a75bdb8247 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.679340] env[62385]: DEBUG oslo_vmware.api [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 942.679340] env[62385]: value = "task-1206153" [ 942.679340] env[62385]: _type = "Task" [ 942.679340] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.686850] env[62385]: DEBUG oslo_vmware.api [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206153, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.796029] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206152, 'name': CreateVM_Task, 'duration_secs': 0.348863} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.796207] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 942.796873] env[62385]: DEBUG oslo_concurrency.lockutils [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.797058] env[62385]: DEBUG oslo_concurrency.lockutils [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.797378] env[62385]: DEBUG oslo_concurrency.lockutils [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 942.797637] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e37690cd-c513-4319-b612-f34d2234474b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.802155] env[62385]: DEBUG oslo_vmware.api [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 942.802155] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]526762bc-b1e2-03f4-7300-cf17f3a609b2" [ 942.802155] env[62385]: _type = "Task" [ 942.802155] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.811611] env[62385]: DEBUG oslo_vmware.api [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]526762bc-b1e2-03f4-7300-cf17f3a609b2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.891140] env[62385]: DEBUG nova.objects.base [None req-f1e1f61c-2527-4118-889d-0942f17b62b3 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62385) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 942.891393] env[62385]: DEBUG nova.network.neutron [None req-f1e1f61c-2527-4118-889d-0942f17b62b3 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 942.998242] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f1e1f61c-2527-4118-889d-0942f17b62b3 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "interface-f937ef7b-ab39-4d8a-9577-7faabc652aa5-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.115s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.076082] env[62385]: DEBUG nova.network.neutron [-] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.092132] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 5daf0a99-0c2a-40d8-afc7-1998e21e32d8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 943.190678] env[62385]: DEBUG oslo_vmware.api [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206153, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.461002} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.193025] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 8e1d416c-2624-468c-94d7-6265e9f4178f/8e1d416c-2624-468c-94d7-6265e9f4178f.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 943.193265] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 943.193524] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-47039793-dbb0-46c9-b742-bcc7d669a265 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.199612] env[62385]: DEBUG oslo_vmware.api [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 943.199612] env[62385]: value = "task-1206154" [ 943.199612] env[62385]: _type = "Task" [ 943.199612] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.207121] env[62385]: DEBUG oslo_vmware.api [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206154, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.312813] env[62385]: DEBUG oslo_vmware.api [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]526762bc-b1e2-03f4-7300-cf17f3a609b2, 'name': SearchDatastore_Task, 'duration_secs': 0.011484} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.313219] env[62385]: DEBUG oslo_concurrency.lockutils [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.313219] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 943.313414] env[62385]: DEBUG oslo_concurrency.lockutils [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.313569] env[62385]: DEBUG oslo_concurrency.lockutils [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.313754] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 943.314034] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-521b4fd3-d05e-431c-b34c-198d269060d9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.322298] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 943.322598] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 943.323246] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27779d09-8a7e-4b95-ad12-1b363a34a0b9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.330097] env[62385]: DEBUG oslo_vmware.api [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 943.330097] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]522e91e1-64e3-4d7f-7130-0220e16938c6" [ 943.330097] env[62385]: _type = "Task" [ 943.330097] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.337819] env[62385]: DEBUG oslo_vmware.api [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]522e91e1-64e3-4d7f-7130-0220e16938c6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.372070] env[62385]: DEBUG nova.network.neutron [req-91be97db-85c3-425f-86d0-97c8e5510af7 req-d32a7cf9-5bef-4e47-893b-51956e4c0907 service nova] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Updated VIF entry in instance network info cache for port 0533b6c1-8d5b-4822-acd6-a95a651ba50c. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 943.372451] env[62385]: DEBUG nova.network.neutron [req-91be97db-85c3-425f-86d0-97c8e5510af7 req-d32a7cf9-5bef-4e47-893b-51956e4c0907 service nova] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Updating instance_info_cache with network_info: [{"id": "0533b6c1-8d5b-4822-acd6-a95a651ba50c", "address": "fa:16:3e:db:e5:07", "network": {"id": "8b2d7a6e-25d4-4a54-ab54-5972612046da", "bridge": "br-int", "label": "tempest-ServersTestJSON-1142134211-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bf8e3c942d3445919cfbe988cca84e90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0533b6c1-8d", "ovs_interfaceid": "0533b6c1-8d5b-4822-acd6-a95a651ba50c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.579538] env[62385]: INFO nova.compute.manager [-] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Took 1.61 seconds to deallocate network for instance. [ 943.595850] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 104dea2c-6977-4683-bc32-05131858f562 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 943.713044] env[62385]: DEBUG oslo_vmware.api [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206154, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061029} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.713044] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 943.713044] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f77ecdcf-b446-4f60-8c76-abef92578726 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.736791] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] 8e1d416c-2624-468c-94d7-6265e9f4178f/8e1d416c-2624-468c-94d7-6265e9f4178f.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 943.736791] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-308ccf4c-e3e6-4391-98ef-02127df529d6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.759513] env[62385]: DEBUG oslo_vmware.api [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 943.759513] env[62385]: value = "task-1206155" [ 943.759513] env[62385]: _type = "Task" [ 943.759513] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.768330] env[62385]: DEBUG oslo_vmware.api [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206155, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.840339] env[62385]: DEBUG oslo_vmware.api [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]522e91e1-64e3-4d7f-7130-0220e16938c6, 'name': SearchDatastore_Task, 'duration_secs': 0.009056} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.841297] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5cc81c68-8c5a-4ef5-89ee-088407557673 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.846987] env[62385]: DEBUG oslo_vmware.api [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 943.846987] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5256cee4-8490-c045-236c-d4ee4fd837a8" [ 943.846987] env[62385]: _type = "Task" [ 943.846987] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.854869] env[62385]: DEBUG oslo_vmware.api [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5256cee4-8490-c045-236c-d4ee4fd837a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.875618] env[62385]: DEBUG oslo_concurrency.lockutils [req-91be97db-85c3-425f-86d0-97c8e5510af7 req-d32a7cf9-5bef-4e47-893b-51956e4c0907 service nova] Releasing lock "refresh_cache-df559cb8-555c-496b-bce3-0981be3ef65f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.090016] env[62385]: DEBUG oslo_concurrency.lockutils [None req-84b85202-2a88-4ae6-a721-84e76137e9f4 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.098812] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance fd0b59c2-e38d-4ecc-b32f-72f07d555ffd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 944.099072] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Total usable vcpus: 48, total allocated vcpus: 11 {{(pid=62385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 944.099222] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2688MB phys_disk=200GB used_disk=11GB total_vcpus=48 used_vcpus=11 pci_stats=[] {{(pid=62385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 944.272523] env[62385]: DEBUG oslo_vmware.api [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206155, 'name': ReconfigVM_Task, 'duration_secs': 0.276062} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.274588] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Reconfigured VM instance instance-0000005b to attach disk [datastore2] 8e1d416c-2624-468c-94d7-6265e9f4178f/8e1d416c-2624-468c-94d7-6265e9f4178f.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 944.275352] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-10236ff1-58f2-4363-934d-f135affb77fa {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.282422] env[62385]: DEBUG oslo_vmware.api [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 944.282422] env[62385]: value = "task-1206156" [ 944.282422] env[62385]: _type = "Task" [ 944.282422] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.297112] env[62385]: DEBUG oslo_vmware.api [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206156, 'name': Rename_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.332963] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09eae080-de9b-497c-bb36-9f45c05f78f9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.340026] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9c3a906-89eb-4d7a-97f1-d39359735dab {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.374086] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1834ca09-41fc-4a64-b370-9dc8ea777840 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.387985] env[62385]: DEBUG oslo_vmware.api [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5256cee4-8490-c045-236c-d4ee4fd837a8, 'name': SearchDatastore_Task, 'duration_secs': 0.009196} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.391503] env[62385]: DEBUG oslo_concurrency.lockutils [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.391503] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] df559cb8-555c-496b-bce3-0981be3ef65f/df559cb8-555c-496b-bce3-0981be3ef65f.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 944.392480] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-20c6b3ad-a63e-4f9d-959c-5aca3be0674c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.395725] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebda72dc-8726-4656-b9bf-fba2df8c1c43 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.407363] env[62385]: DEBUG oslo_vmware.api [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 944.407363] env[62385]: value = "task-1206157" [ 944.407363] env[62385]: _type = "Task" [ 944.407363] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.417034] env[62385]: DEBUG nova.compute.provider_tree [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 944.426975] env[62385]: DEBUG oslo_vmware.api [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206157, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.795350] env[62385]: DEBUG oslo_vmware.api [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206156, 'name': Rename_Task, 'duration_secs': 0.140146} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.795665] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 944.795981] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dbc1f96c-1bdd-415b-88db-524992e81528 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.804865] env[62385]: DEBUG oslo_vmware.api [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 944.804865] env[62385]: value = "task-1206158" [ 944.804865] env[62385]: _type = "Task" [ 944.804865] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.813794] env[62385]: DEBUG oslo_vmware.api [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206158, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.883175] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e6dd994d-385e-4b5b-a2e3-085be93a7d91 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "a58cb15f-e4df-4b1e-b09c-2f61dce7200f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.883175] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e6dd994d-385e-4b5b-a2e3-085be93a7d91 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "a58cb15f-e4df-4b1e-b09c-2f61dce7200f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.883175] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e6dd994d-385e-4b5b-a2e3-085be93a7d91 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "a58cb15f-e4df-4b1e-b09c-2f61dce7200f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.883554] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e6dd994d-385e-4b5b-a2e3-085be93a7d91 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "a58cb15f-e4df-4b1e-b09c-2f61dce7200f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.883554] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e6dd994d-385e-4b5b-a2e3-085be93a7d91 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "a58cb15f-e4df-4b1e-b09c-2f61dce7200f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.889604] env[62385]: INFO nova.compute.manager [None req-e6dd994d-385e-4b5b-a2e3-085be93a7d91 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Terminating instance [ 944.895796] env[62385]: DEBUG nova.compute.manager [None req-e6dd994d-385e-4b5b-a2e3-085be93a7d91 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 944.896010] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-e6dd994d-385e-4b5b-a2e3-085be93a7d91 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 944.896944] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21506a09-107f-4480-bfdc-e86b4d3c9fba {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.906621] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6dd994d-385e-4b5b-a2e3-085be93a7d91 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 944.906621] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fd56a7e7-c814-4c24-9744-bdda6425ab5b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.913328] env[62385]: DEBUG oslo_vmware.api [None req-e6dd994d-385e-4b5b-a2e3-085be93a7d91 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 944.913328] env[62385]: value = "task-1206159" [ 944.913328] env[62385]: _type = "Task" [ 944.913328] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.926131] env[62385]: DEBUG nova.scheduler.client.report [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 944.929848] env[62385]: DEBUG oslo_vmware.api [None req-e6dd994d-385e-4b5b-a2e3-085be93a7d91 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206159, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.934108] env[62385]: DEBUG oslo_vmware.api [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206157, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.444908} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.934108] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] df559cb8-555c-496b-bce3-0981be3ef65f/df559cb8-555c-496b-bce3-0981be3ef65f.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 944.934341] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 944.934624] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3a681723-5dd9-45ae-bf05-d71b31a3cddc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.941505] env[62385]: DEBUG oslo_vmware.api [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 944.941505] env[62385]: value = "task-1206160" [ 944.941505] env[62385]: _type = "Task" [ 944.941505] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.952931] env[62385]: DEBUG oslo_vmware.api [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206160, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.975663] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a86e2e6f-9cfc-4d35-b6f5-b06a789b3209 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "interface-f937ef7b-ab39-4d8a-9577-7faabc652aa5-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.976148] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a86e2e6f-9cfc-4d35-b6f5-b06a789b3209 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "interface-f937ef7b-ab39-4d8a-9577-7faabc652aa5-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.976753] env[62385]: DEBUG nova.objects.instance [None req-a86e2e6f-9cfc-4d35-b6f5-b06a789b3209 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lazy-loading 'flavor' on Instance uuid f937ef7b-ab39-4d8a-9577-7faabc652aa5 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 944.980473] env[62385]: DEBUG nova.compute.manager [req-2c4efd3e-558e-4c8b-b6c9-a49803d6aef2 req-2c7511e8-2911-454f-bc3e-0c38da799619 service nova] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Received event network-vif-deleted-89da5ce4-4773-4c11-91a4-870768a5b165 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 944.981093] env[62385]: DEBUG nova.compute.manager [req-2c4efd3e-558e-4c8b-b6c9-a49803d6aef2 req-2c7511e8-2911-454f-bc3e-0c38da799619 service nova] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Received event network-vif-deleted-e2a123f6-199e-4e54-bd44-8b64435eebca {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 945.316670] env[62385]: DEBUG oslo_vmware.api [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206158, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.385546] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "df020d6b-3fab-4599-a342-47c7833b4240" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.385799] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "df020d6b-3fab-4599-a342-47c7833b4240" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.424131] env[62385]: DEBUG oslo_vmware.api [None req-e6dd994d-385e-4b5b-a2e3-085be93a7d91 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206159, 'name': PowerOffVM_Task, 'duration_secs': 0.206875} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.424368] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6dd994d-385e-4b5b-a2e3-085be93a7d91 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 945.424542] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-e6dd994d-385e-4b5b-a2e3-085be93a7d91 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 945.424831] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8384484e-e512-49e6-960c-b9fd96be1fd8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.434198] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62385) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 945.434431] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.889s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.434745] env[62385]: DEBUG oslo_concurrency.lockutils [None req-cb13311b-0627-4f9e-94c6-e089812528d0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.292s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.434975] env[62385]: DEBUG oslo_concurrency.lockutils [None req-cb13311b-0627-4f9e-94c6-e089812528d0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.436965] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.754s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.438610] env[62385]: INFO nova.compute.claims [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 945.451885] env[62385]: DEBUG oslo_vmware.api [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206160, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061724} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.452295] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 945.453320] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dab3470-2d79-4f56-8583-f67efb67b9fb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.478491] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] df559cb8-555c-496b-bce3-0981be3ef65f/df559cb8-555c-496b-bce3-0981be3ef65f.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 945.479782] env[62385]: INFO nova.scheduler.client.report [None req-cb13311b-0627-4f9e-94c6-e089812528d0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Deleted allocations for instance 679150f5-ad17-45e6-9e3d-9f6e28b1aee3 [ 945.481130] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-473ffb68-5865-4a32-9a08-88aa94bbd2a9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.505672] env[62385]: DEBUG oslo_vmware.api [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 945.505672] env[62385]: value = "task-1206162" [ 945.505672] env[62385]: _type = "Task" [ 945.505672] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.512430] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-e6dd994d-385e-4b5b-a2e3-085be93a7d91 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 945.512686] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-e6dd994d-385e-4b5b-a2e3-085be93a7d91 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 945.512828] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6dd994d-385e-4b5b-a2e3-085be93a7d91 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Deleting the datastore file [datastore1] a58cb15f-e4df-4b1e-b09c-2f61dce7200f {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 945.516583] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bbe72d49-a83d-48a0-b784-0c92a6531cee {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.518884] env[62385]: DEBUG oslo_vmware.api [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206162, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.524574] env[62385]: DEBUG oslo_vmware.api [None req-e6dd994d-385e-4b5b-a2e3-085be93a7d91 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 945.524574] env[62385]: value = "task-1206163" [ 945.524574] env[62385]: _type = "Task" [ 945.524574] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.533764] env[62385]: DEBUG oslo_vmware.api [None req-e6dd994d-385e-4b5b-a2e3-085be93a7d91 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206163, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.621256] env[62385]: DEBUG nova.objects.instance [None req-a86e2e6f-9cfc-4d35-b6f5-b06a789b3209 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lazy-loading 'pci_requests' on Instance uuid f937ef7b-ab39-4d8a-9577-7faabc652aa5 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 945.815367] env[62385]: DEBUG oslo_vmware.api [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206158, 'name': PowerOnVM_Task, 'duration_secs': 0.722493} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.815756] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 945.815942] env[62385]: INFO nova.compute.manager [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Took 8.91 seconds to spawn the instance on the hypervisor. [ 945.816155] env[62385]: DEBUG nova.compute.manager [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 945.816931] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ddbacf6-b8c1-4c3d-aefb-1f1add7f2e05 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.888373] env[62385]: DEBUG nova.compute.manager [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 946.003603] env[62385]: DEBUG oslo_concurrency.lockutils [None req-cb13311b-0627-4f9e-94c6-e089812528d0 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "679150f5-ad17-45e6-9e3d-9f6e28b1aee3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.893s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.014942] env[62385]: DEBUG oslo_vmware.api [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206162, 'name': ReconfigVM_Task, 'duration_secs': 0.278899} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.015256] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Reconfigured VM instance instance-0000005c to attach disk [datastore1] df559cb8-555c-496b-bce3-0981be3ef65f/df559cb8-555c-496b-bce3-0981be3ef65f.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 946.015919] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-88381355-5dde-485a-9f1e-f04afcf30222 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.023302] env[62385]: DEBUG oslo_vmware.api [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 946.023302] env[62385]: value = "task-1206164" [ 946.023302] env[62385]: _type = "Task" [ 946.023302] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.033932] env[62385]: DEBUG oslo_vmware.api [None req-e6dd994d-385e-4b5b-a2e3-085be93a7d91 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206163, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154658} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.036833] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6dd994d-385e-4b5b-a2e3-085be93a7d91 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 946.037049] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-e6dd994d-385e-4b5b-a2e3-085be93a7d91 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 946.037240] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-e6dd994d-385e-4b5b-a2e3-085be93a7d91 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 946.037423] env[62385]: INFO nova.compute.manager [None req-e6dd994d-385e-4b5b-a2e3-085be93a7d91 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Took 1.14 seconds to destroy the instance on the hypervisor. [ 946.037659] env[62385]: DEBUG oslo.service.loopingcall [None req-e6dd994d-385e-4b5b-a2e3-085be93a7d91 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 946.038103] env[62385]: DEBUG oslo_vmware.api [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206164, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.038319] env[62385]: DEBUG nova.compute.manager [-] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 946.038415] env[62385]: DEBUG nova.network.neutron [-] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 946.123855] env[62385]: DEBUG nova.objects.base [None req-a86e2e6f-9cfc-4d35-b6f5-b06a789b3209 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62385) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 946.124105] env[62385]: DEBUG nova.network.neutron [None req-a86e2e6f-9cfc-4d35-b6f5-b06a789b3209 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 946.224787] env[62385]: DEBUG nova.policy [None req-a86e2e6f-9cfc-4d35-b6f5-b06a789b3209 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5040b20c848c4d28a198f1773a3ff21e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '012426f098ce40c3aaa00f628fe9cebb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 946.334786] env[62385]: INFO nova.compute.manager [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Took 29.46 seconds to build instance. [ 946.411655] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.535673] env[62385]: DEBUG oslo_vmware.api [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206164, 'name': Rename_Task, 'duration_secs': 0.246002} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.538647] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 946.539191] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a8aae934-3bc6-4193-82d1-22383d8004a3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.547025] env[62385]: DEBUG oslo_vmware.api [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 946.547025] env[62385]: value = "task-1206165" [ 946.547025] env[62385]: _type = "Task" [ 946.547025] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.559198] env[62385]: DEBUG oslo_vmware.api [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206165, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.613031] env[62385]: DEBUG nova.network.neutron [None req-a86e2e6f-9cfc-4d35-b6f5-b06a789b3209 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Successfully created port: c053ce8b-fc9f-4c48-9073-cdd753630700 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 946.718360] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a943bc0a-4c82-487d-9ca8-7cfc9e804b02 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.726051] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1444a38-3eef-4cae-aac0-a5b0e17dbe34 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.760233] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3d59e1f-2cff-48ee-a4c0-145fada9c4f2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.768366] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-750e27bc-f698-4435-aa63-a24c9b81f6a3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.782475] env[62385]: DEBUG nova.compute.provider_tree [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 946.836533] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ba8b3202-e4f1-4a43-ba2a-900c29d15d87 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lock "8e1d416c-2624-468c-94d7-6265e9f4178f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.970s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.900031] env[62385]: DEBUG nova.network.neutron [-] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.058148] env[62385]: DEBUG oslo_vmware.api [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206165, 'name': PowerOnVM_Task, 'duration_secs': 0.460527} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.058412] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 947.058615] env[62385]: INFO nova.compute.manager [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Took 7.76 seconds to spawn the instance on the hypervisor. [ 947.058798] env[62385]: DEBUG nova.compute.manager [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 947.059591] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc3d8f7a-c2c2-436b-b379-755f7943613a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.136824] env[62385]: DEBUG nova.compute.manager [req-07908548-20ff-411b-b5f2-ce00905be232 req-1c4fbbc2-03dc-4755-a9e6-d274fe97b358 service nova] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Received event network-vif-deleted-f5cfcc74-0f00-411e-8ee7-e26ec0a2fe7a {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 947.170320] env[62385]: DEBUG oslo_concurrency.lockutils [None req-782e1513-c3dc-4b59-8b0f-73e617258048 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquiring lock "f1d33401-35f5-4d79-abb4-26dc6faa784e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.170632] env[62385]: DEBUG oslo_concurrency.lockutils [None req-782e1513-c3dc-4b59-8b0f-73e617258048 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "f1d33401-35f5-4d79-abb4-26dc6faa784e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.170873] env[62385]: DEBUG oslo_concurrency.lockutils [None req-782e1513-c3dc-4b59-8b0f-73e617258048 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquiring lock "f1d33401-35f5-4d79-abb4-26dc6faa784e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.171075] env[62385]: DEBUG oslo_concurrency.lockutils [None req-782e1513-c3dc-4b59-8b0f-73e617258048 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "f1d33401-35f5-4d79-abb4-26dc6faa784e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.171270] env[62385]: DEBUG oslo_concurrency.lockutils [None req-782e1513-c3dc-4b59-8b0f-73e617258048 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "f1d33401-35f5-4d79-abb4-26dc6faa784e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.173305] env[62385]: INFO nova.compute.manager [None req-782e1513-c3dc-4b59-8b0f-73e617258048 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Terminating instance [ 947.175383] env[62385]: DEBUG nova.compute.manager [None req-782e1513-c3dc-4b59-8b0f-73e617258048 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 947.175383] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-782e1513-c3dc-4b59-8b0f-73e617258048 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 947.176433] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8296468d-e363-43ce-a00c-15db5e116144 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.185319] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-782e1513-c3dc-4b59-8b0f-73e617258048 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 947.185567] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6efc34f7-85d9-4d7a-b16e-c938b16b1d97 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.191867] env[62385]: DEBUG oslo_vmware.api [None req-782e1513-c3dc-4b59-8b0f-73e617258048 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 947.191867] env[62385]: value = "task-1206166" [ 947.191867] env[62385]: _type = "Task" [ 947.191867] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.199906] env[62385]: DEBUG oslo_vmware.api [None req-782e1513-c3dc-4b59-8b0f-73e617258048 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206166, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.218939] env[62385]: DEBUG nova.compute.manager [req-f3adfd49-8cc7-4d03-9059-0c1ee76a21d0 req-e718077d-2e5b-4ae1-a098-b6ef5829f22c service nova] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Received event network-changed-deafdeb9-f59c-48f4-873b-6e4b27d4a688 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 947.218939] env[62385]: DEBUG nova.compute.manager [req-f3adfd49-8cc7-4d03-9059-0c1ee76a21d0 req-e718077d-2e5b-4ae1-a098-b6ef5829f22c service nova] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Refreshing instance network info cache due to event network-changed-deafdeb9-f59c-48f4-873b-6e4b27d4a688. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 947.219354] env[62385]: DEBUG oslo_concurrency.lockutils [req-f3adfd49-8cc7-4d03-9059-0c1ee76a21d0 req-e718077d-2e5b-4ae1-a098-b6ef5829f22c service nova] Acquiring lock "refresh_cache-8e1d416c-2624-468c-94d7-6265e9f4178f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.219615] env[62385]: DEBUG oslo_concurrency.lockutils [req-f3adfd49-8cc7-4d03-9059-0c1ee76a21d0 req-e718077d-2e5b-4ae1-a098-b6ef5829f22c service nova] Acquired lock "refresh_cache-8e1d416c-2624-468c-94d7-6265e9f4178f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.219814] env[62385]: DEBUG nova.network.neutron [req-f3adfd49-8cc7-4d03-9059-0c1ee76a21d0 req-e718077d-2e5b-4ae1-a098-b6ef5829f22c service nova] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Refreshing network info cache for port deafdeb9-f59c-48f4-873b-6e4b27d4a688 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 947.285240] env[62385]: DEBUG nova.scheduler.client.report [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 947.404999] env[62385]: INFO nova.compute.manager [-] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Took 1.37 seconds to deallocate network for instance. [ 947.582712] env[62385]: INFO nova.compute.manager [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Took 30.07 seconds to build instance. [ 947.702700] env[62385]: DEBUG oslo_vmware.api [None req-782e1513-c3dc-4b59-8b0f-73e617258048 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206166, 'name': PowerOffVM_Task, 'duration_secs': 0.375132} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.702965] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-782e1513-c3dc-4b59-8b0f-73e617258048 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 947.703160] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-782e1513-c3dc-4b59-8b0f-73e617258048 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 947.703429] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3ac31020-de9c-49f3-a6e1-7235a0a8eb25 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.767067] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-782e1513-c3dc-4b59-8b0f-73e617258048 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 947.767383] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-782e1513-c3dc-4b59-8b0f-73e617258048 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Deleting contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 947.767782] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-782e1513-c3dc-4b59-8b0f-73e617258048 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Deleting the datastore file [datastore2] f1d33401-35f5-4d79-abb4-26dc6faa784e {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 947.768048] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d6778a7e-205e-42b7-9df6-95b2b74fc8c3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.774842] env[62385]: DEBUG oslo_vmware.api [None req-782e1513-c3dc-4b59-8b0f-73e617258048 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 947.774842] env[62385]: value = "task-1206168" [ 947.774842] env[62385]: _type = "Task" [ 947.774842] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.784502] env[62385]: DEBUG oslo_vmware.api [None req-782e1513-c3dc-4b59-8b0f-73e617258048 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206168, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.978861] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.355s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.978861] env[62385]: DEBUG nova.compute.manager [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 947.978861] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0d3ddba5-96d8-4d1f-a0cb-bb6183881104 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.149s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.978861] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0d3ddba5-96d8-4d1f-a0cb-bb6183881104 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.978861] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.869s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.978861] env[62385]: INFO nova.compute.claims [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] [instance: 104dea2c-6977-4683-bc32-05131858f562] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 947.978861] env[62385]: INFO nova.scheduler.client.report [None req-0d3ddba5-96d8-4d1f-a0cb-bb6183881104 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Deleted allocations for instance fad3b054-cf2a-4ee3-a2a5-918a96a4a125 [ 947.978861] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e6dd994d-385e-4b5b-a2e3-085be93a7d91 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.084648] env[62385]: DEBUG oslo_concurrency.lockutils [None req-77e09bd0-4191-4bd6-95ac-49d8f67f091b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "df559cb8-555c-496b-bce3-0981be3ef65f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.595s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.187834] env[62385]: DEBUG nova.network.neutron [req-f3adfd49-8cc7-4d03-9059-0c1ee76a21d0 req-e718077d-2e5b-4ae1-a098-b6ef5829f22c service nova] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Updated VIF entry in instance network info cache for port deafdeb9-f59c-48f4-873b-6e4b27d4a688. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 948.188228] env[62385]: DEBUG nova.network.neutron [req-f3adfd49-8cc7-4d03-9059-0c1ee76a21d0 req-e718077d-2e5b-4ae1-a098-b6ef5829f22c service nova] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Updating instance_info_cache with network_info: [{"id": "deafdeb9-f59c-48f4-873b-6e4b27d4a688", "address": "fa:16:3e:81:4d:75", "network": {"id": "9c4d6464-cacc-44f8-b7e5-0699cd312c53", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1889153615-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.186", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09c2c5e938bf46fd9353e47aecc72478", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4d3f69a-b086-4c3b-b976-5a848b63dfc4", "external-id": "nsx-vlan-transportzone-627", "segmentation_id": 627, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdeafdeb9-f5", "ovs_interfaceid": "deafdeb9-f59c-48f4-873b-6e4b27d4a688", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.284793] env[62385]: DEBUG oslo_vmware.api [None req-782e1513-c3dc-4b59-8b0f-73e617258048 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206168, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145678} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.285075] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-782e1513-c3dc-4b59-8b0f-73e617258048 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 948.285289] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-782e1513-c3dc-4b59-8b0f-73e617258048 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Deleted contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 948.285543] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-782e1513-c3dc-4b59-8b0f-73e617258048 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 948.285715] env[62385]: INFO nova.compute.manager [None req-782e1513-c3dc-4b59-8b0f-73e617258048 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Took 1.11 seconds to destroy the instance on the hypervisor. [ 948.285947] env[62385]: DEBUG oslo.service.loopingcall [None req-782e1513-c3dc-4b59-8b0f-73e617258048 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 948.286315] env[62385]: DEBUG nova.compute.manager [-] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 948.286315] env[62385]: DEBUG nova.network.neutron [-] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 948.304444] env[62385]: DEBUG nova.compute.utils [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 948.309028] env[62385]: DEBUG nova.compute.manager [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 948.309028] env[62385]: DEBUG nova.network.neutron [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 948.352364] env[62385]: DEBUG nova.policy [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b314c392b9ca405480c27c25364fda81', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '93111f5ba17c47de9691f90b99e03aa8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 948.357806] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0d3ddba5-96d8-4d1f-a0cb-bb6183881104 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "fad3b054-cf2a-4ee3-a2a5-918a96a4a125" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.466s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.572711] env[62385]: DEBUG nova.network.neutron [None req-a86e2e6f-9cfc-4d35-b6f5-b06a789b3209 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Successfully updated port: c053ce8b-fc9f-4c48-9073-cdd753630700 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 948.668151] env[62385]: DEBUG nova.network.neutron [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Successfully created port: 1b04305d-1213-4f55-9b69-6f15de71f576 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 948.690657] env[62385]: DEBUG oslo_concurrency.lockutils [req-f3adfd49-8cc7-4d03-9059-0c1ee76a21d0 req-e718077d-2e5b-4ae1-a098-b6ef5829f22c service nova] Releasing lock "refresh_cache-8e1d416c-2624-468c-94d7-6265e9f4178f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.811384] env[62385]: DEBUG nova.compute.manager [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 949.033353] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e68075e-311e-4a6e-9e74-8aa8a006f82b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.041162] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf020417-79c4-4cee-83ec-2efb2ec5f5b2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.071393] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f9561cd-4e17-48f0-badb-410d2c611819 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.074251] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a86e2e6f-9cfc-4d35-b6f5-b06a789b3209 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "refresh_cache-f937ef7b-ab39-4d8a-9577-7faabc652aa5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.074422] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a86e2e6f-9cfc-4d35-b6f5-b06a789b3209 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquired lock "refresh_cache-f937ef7b-ab39-4d8a-9577-7faabc652aa5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.074601] env[62385]: DEBUG nova.network.neutron [None req-a86e2e6f-9cfc-4d35-b6f5-b06a789b3209 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 949.080840] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-807315dd-c592-4c3f-b478-4b2e04b81991 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.096128] env[62385]: DEBUG nova.compute.provider_tree [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 949.178392] env[62385]: DEBUG nova.network.neutron [-] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 949.954686] env[62385]: DEBUG nova.scheduler.client.report [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 949.957515] env[62385]: INFO nova.compute.manager [-] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Took 1.67 seconds to deallocate network for instance. [ 949.960564] env[62385]: DEBUG nova.compute.manager [req-272dfbda-e6dc-4b52-aa97-737ba96a3d84 req-b7e1c8db-5960-46a5-8c8a-11c4f637657a service nova] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Received event network-vif-plugged-c053ce8b-fc9f-4c48-9073-cdd753630700 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 949.961422] env[62385]: DEBUG oslo_concurrency.lockutils [req-272dfbda-e6dc-4b52-aa97-737ba96a3d84 req-b7e1c8db-5960-46a5-8c8a-11c4f637657a service nova] Acquiring lock "f937ef7b-ab39-4d8a-9577-7faabc652aa5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.961422] env[62385]: DEBUG oslo_concurrency.lockutils [req-272dfbda-e6dc-4b52-aa97-737ba96a3d84 req-b7e1c8db-5960-46a5-8c8a-11c4f637657a service nova] Lock "f937ef7b-ab39-4d8a-9577-7faabc652aa5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.961422] env[62385]: DEBUG oslo_concurrency.lockutils [req-272dfbda-e6dc-4b52-aa97-737ba96a3d84 req-b7e1c8db-5960-46a5-8c8a-11c4f637657a service nova] Lock "f937ef7b-ab39-4d8a-9577-7faabc652aa5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.961422] env[62385]: DEBUG nova.compute.manager [req-272dfbda-e6dc-4b52-aa97-737ba96a3d84 req-b7e1c8db-5960-46a5-8c8a-11c4f637657a service nova] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] No waiting events found dispatching network-vif-plugged-c053ce8b-fc9f-4c48-9073-cdd753630700 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 949.961627] env[62385]: WARNING nova.compute.manager [req-272dfbda-e6dc-4b52-aa97-737ba96a3d84 req-b7e1c8db-5960-46a5-8c8a-11c4f637657a service nova] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Received unexpected event network-vif-plugged-c053ce8b-fc9f-4c48-9073-cdd753630700 for instance with vm_state active and task_state None. [ 949.963249] env[62385]: DEBUG nova.compute.manager [req-272dfbda-e6dc-4b52-aa97-737ba96a3d84 req-b7e1c8db-5960-46a5-8c8a-11c4f637657a service nova] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Received event network-changed-c053ce8b-fc9f-4c48-9073-cdd753630700 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 949.963249] env[62385]: DEBUG nova.compute.manager [req-272dfbda-e6dc-4b52-aa97-737ba96a3d84 req-b7e1c8db-5960-46a5-8c8a-11c4f637657a service nova] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Refreshing instance network info cache due to event network-changed-c053ce8b-fc9f-4c48-9073-cdd753630700. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 949.963249] env[62385]: DEBUG oslo_concurrency.lockutils [req-272dfbda-e6dc-4b52-aa97-737ba96a3d84 req-b7e1c8db-5960-46a5-8c8a-11c4f637657a service nova] Acquiring lock "refresh_cache-f937ef7b-ab39-4d8a-9577-7faabc652aa5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.966385] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "b1e0baf3-643c-49c2-8a80-8de07d8527e7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.966587] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "b1e0baf3-643c-49c2-8a80-8de07d8527e7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.000825] env[62385]: WARNING nova.network.neutron [None req-a86e2e6f-9cfc-4d35-b6f5-b06a789b3209 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] daad7070-dc06-4580-9f4a-ac8aaa546955 already exists in list: networks containing: ['daad7070-dc06-4580-9f4a-ac8aaa546955']. ignoring it [ 950.140239] env[62385]: DEBUG nova.compute.manager [req-3aeb56b1-eaf7-4110-ae87-e53e697414ab req-74db3536-d446-4c10-b3d2-a60fa5ab02e3 service nova] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Received event network-vif-plugged-1b04305d-1213-4f55-9b69-6f15de71f576 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 950.140534] env[62385]: DEBUG oslo_concurrency.lockutils [req-3aeb56b1-eaf7-4110-ae87-e53e697414ab req-74db3536-d446-4c10-b3d2-a60fa5ab02e3 service nova] Acquiring lock "5daf0a99-0c2a-40d8-afc7-1998e21e32d8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.140779] env[62385]: DEBUG oslo_concurrency.lockutils [req-3aeb56b1-eaf7-4110-ae87-e53e697414ab req-74db3536-d446-4c10-b3d2-a60fa5ab02e3 service nova] Lock "5daf0a99-0c2a-40d8-afc7-1998e21e32d8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.140985] env[62385]: DEBUG oslo_concurrency.lockutils [req-3aeb56b1-eaf7-4110-ae87-e53e697414ab req-74db3536-d446-4c10-b3d2-a60fa5ab02e3 service nova] Lock "5daf0a99-0c2a-40d8-afc7-1998e21e32d8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.141086] env[62385]: DEBUG nova.compute.manager [req-3aeb56b1-eaf7-4110-ae87-e53e697414ab req-74db3536-d446-4c10-b3d2-a60fa5ab02e3 service nova] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] No waiting events found dispatching network-vif-plugged-1b04305d-1213-4f55-9b69-6f15de71f576 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 950.141224] env[62385]: WARNING nova.compute.manager [req-3aeb56b1-eaf7-4110-ae87-e53e697414ab req-74db3536-d446-4c10-b3d2-a60fa5ab02e3 service nova] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Received unexpected event network-vif-plugged-1b04305d-1213-4f55-9b69-6f15de71f576 for instance with vm_state building and task_state spawning. [ 950.190117] env[62385]: DEBUG nova.network.neutron [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Successfully updated port: 1b04305d-1213-4f55-9b69-6f15de71f576 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 950.342059] env[62385]: DEBUG nova.network.neutron [None req-a86e2e6f-9cfc-4d35-b6f5-b06a789b3209 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Updating instance_info_cache with network_info: [{"id": "2b05b2b6-d4ce-4eb7-9d65-730d4dbad51c", "address": "fa:16:3e:ec:d7:16", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b05b2b6-d4", "ovs_interfaceid": "2b05b2b6-d4ce-4eb7-9d65-730d4dbad51c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c053ce8b-fc9f-4c48-9073-cdd753630700", "address": "fa:16:3e:9c:f6:87", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc053ce8b-fc", "ovs_interfaceid": "c053ce8b-fc9f-4c48-9073-cdd753630700", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.465809] env[62385]: DEBUG nova.compute.manager [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 950.468529] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c319277e-63ff-46fd-a975-49b0644f34a8 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "df559cb8-555c-496b-bce3-0981be3ef65f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.468806] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c319277e-63ff-46fd-a975-49b0644f34a8 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "df559cb8-555c-496b-bce3-0981be3ef65f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.469058] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c319277e-63ff-46fd-a975-49b0644f34a8 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "df559cb8-555c-496b-bce3-0981be3ef65f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.469258] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c319277e-63ff-46fd-a975-49b0644f34a8 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "df559cb8-555c-496b-bce3-0981be3ef65f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.469429] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c319277e-63ff-46fd-a975-49b0644f34a8 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "df559cb8-555c-496b-bce3-0981be3ef65f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.471571] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.673s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.472043] env[62385]: DEBUG nova.compute.manager [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] [instance: 104dea2c-6977-4683-bc32-05131858f562] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 950.474762] env[62385]: DEBUG oslo_concurrency.lockutils [None req-782e1513-c3dc-4b59-8b0f-73e617258048 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.475072] env[62385]: DEBUG nova.compute.manager [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 950.477658] env[62385]: INFO nova.compute.manager [None req-c319277e-63ff-46fd-a975-49b0644f34a8 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Terminating instance [ 950.478857] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e75e6879-4fef-42a9-ba97-4fe18524ff8a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.201s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.478953] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e75e6879-4fef-42a9-ba97-4fe18524ff8a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.480896] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fea8ac26-d2cb-41ad-8f71-07e51195c9f1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.702s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.482008] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fea8ac26-d2cb-41ad-8f71-07e51195c9f1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.482852] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.729s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.484215] env[62385]: INFO nova.compute.claims [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 950.487019] env[62385]: DEBUG nova.compute.manager [None req-c319277e-63ff-46fd-a975-49b0644f34a8 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 950.487215] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c319277e-63ff-46fd-a975-49b0644f34a8 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 950.488917] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89499763-825f-45df-a759-17bdb8dd3b07 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.497049] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c319277e-63ff-46fd-a975-49b0644f34a8 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 950.498666] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5efdb36b-f703-4b0e-b767-c12053ff65b4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.506456] env[62385]: DEBUG oslo_vmware.api [None req-c319277e-63ff-46fd-a975-49b0644f34a8 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 950.506456] env[62385]: value = "task-1206169" [ 950.506456] env[62385]: _type = "Task" [ 950.506456] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.508678] env[62385]: DEBUG nova.virt.hardware [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 950.509000] env[62385]: DEBUG nova.virt.hardware [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 950.509103] env[62385]: DEBUG nova.virt.hardware [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 950.509306] env[62385]: DEBUG nova.virt.hardware [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 950.509858] env[62385]: DEBUG nova.virt.hardware [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 950.509858] env[62385]: DEBUG nova.virt.hardware [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 950.509858] env[62385]: DEBUG nova.virt.hardware [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 950.510065] env[62385]: DEBUG nova.virt.hardware [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 950.510203] env[62385]: DEBUG nova.virt.hardware [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 950.510376] env[62385]: DEBUG nova.virt.hardware [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 950.510602] env[62385]: DEBUG nova.virt.hardware [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 950.511501] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fdc10bd-0521-4561-99e1-acfccfdf35b0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.516267] env[62385]: INFO nova.scheduler.client.report [None req-e75e6879-4fef-42a9-ba97-4fe18524ff8a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Deleted allocations for instance caab700b-bd3c-497e-b40a-eb9d20dfc1e2 [ 950.518882] env[62385]: INFO nova.scheduler.client.report [None req-fea8ac26-d2cb-41ad-8f71-07e51195c9f1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Deleted allocations for instance d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b [ 950.533658] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1d9bbbc-d485-4b3e-96ce-5460142c41b6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.540584] env[62385]: DEBUG oslo_vmware.api [None req-c319277e-63ff-46fd-a975-49b0644f34a8 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206169, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.692553] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "refresh_cache-5daf0a99-0c2a-40d8-afc7-1998e21e32d8" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.692753] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquired lock "refresh_cache-5daf0a99-0c2a-40d8-afc7-1998e21e32d8" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.692917] env[62385]: DEBUG nova.network.neutron [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 950.844841] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a86e2e6f-9cfc-4d35-b6f5-b06a789b3209 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Releasing lock "refresh_cache-f937ef7b-ab39-4d8a-9577-7faabc652aa5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.845524] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a86e2e6f-9cfc-4d35-b6f5-b06a789b3209 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "f937ef7b-ab39-4d8a-9577-7faabc652aa5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.845694] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a86e2e6f-9cfc-4d35-b6f5-b06a789b3209 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquired lock "f937ef7b-ab39-4d8a-9577-7faabc652aa5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.846013] env[62385]: DEBUG oslo_concurrency.lockutils [req-272dfbda-e6dc-4b52-aa97-737ba96a3d84 req-b7e1c8db-5960-46a5-8c8a-11c4f637657a service nova] Acquired lock "refresh_cache-f937ef7b-ab39-4d8a-9577-7faabc652aa5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.846239] env[62385]: DEBUG nova.network.neutron [req-272dfbda-e6dc-4b52-aa97-737ba96a3d84 req-b7e1c8db-5960-46a5-8c8a-11c4f637657a service nova] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Refreshing network info cache for port c053ce8b-fc9f-4c48-9073-cdd753630700 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 950.847980] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b01a93a-f556-47fe-ae71-9eb789b66668 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.865371] env[62385]: DEBUG nova.virt.hardware [None req-a86e2e6f-9cfc-4d35-b6f5-b06a789b3209 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 950.865686] env[62385]: DEBUG nova.virt.hardware [None req-a86e2e6f-9cfc-4d35-b6f5-b06a789b3209 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 950.865988] env[62385]: DEBUG nova.virt.hardware [None req-a86e2e6f-9cfc-4d35-b6f5-b06a789b3209 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 950.866269] env[62385]: DEBUG nova.virt.hardware [None req-a86e2e6f-9cfc-4d35-b6f5-b06a789b3209 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 950.866496] env[62385]: DEBUG nova.virt.hardware [None req-a86e2e6f-9cfc-4d35-b6f5-b06a789b3209 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 950.866667] env[62385]: DEBUG nova.virt.hardware [None req-a86e2e6f-9cfc-4d35-b6f5-b06a789b3209 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 950.866944] env[62385]: DEBUG nova.virt.hardware [None req-a86e2e6f-9cfc-4d35-b6f5-b06a789b3209 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 950.867151] env[62385]: DEBUG nova.virt.hardware [None req-a86e2e6f-9cfc-4d35-b6f5-b06a789b3209 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 950.867339] env[62385]: DEBUG nova.virt.hardware [None req-a86e2e6f-9cfc-4d35-b6f5-b06a789b3209 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 950.867554] env[62385]: DEBUG nova.virt.hardware [None req-a86e2e6f-9cfc-4d35-b6f5-b06a789b3209 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 950.867743] env[62385]: DEBUG nova.virt.hardware [None req-a86e2e6f-9cfc-4d35-b6f5-b06a789b3209 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 950.874294] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a86e2e6f-9cfc-4d35-b6f5-b06a789b3209 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Reconfiguring VM to attach interface {{(pid=62385) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 950.875464] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b8e893f7-f6b2-491e-93cc-425eb13a1a8b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.892524] env[62385]: DEBUG oslo_vmware.api [None req-a86e2e6f-9cfc-4d35-b6f5-b06a789b3209 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 950.892524] env[62385]: value = "task-1206170" [ 950.892524] env[62385]: _type = "Task" [ 950.892524] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.901174] env[62385]: DEBUG oslo_vmware.api [None req-a86e2e6f-9cfc-4d35-b6f5-b06a789b3209 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206170, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.979406] env[62385]: DEBUG nova.compute.utils [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 950.980933] env[62385]: DEBUG nova.compute.manager [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] [instance: 104dea2c-6977-4683-bc32-05131858f562] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 950.981147] env[62385]: DEBUG nova.network.neutron [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] [instance: 104dea2c-6977-4683-bc32-05131858f562] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 951.007082] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.021691] env[62385]: DEBUG oslo_vmware.api [None req-c319277e-63ff-46fd-a975-49b0644f34a8 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206169, 'name': PowerOffVM_Task, 'duration_secs': 0.406642} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.021992] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c319277e-63ff-46fd-a975-49b0644f34a8 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 951.022661] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c319277e-63ff-46fd-a975-49b0644f34a8 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 951.022661] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9b7e3957-dc76-483c-ac3d-b00305a41195 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.031938] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fea8ac26-d2cb-41ad-8f71-07e51195c9f1 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.802s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.034067] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e75e6879-4fef-42a9-ba97-4fe18524ff8a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "caab700b-bd3c-497e-b40a-eb9d20dfc1e2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.495s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.037496] env[62385]: DEBUG nova.policy [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '585202de0a0343c78fa39a61e0c6ba28', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2a9f68fd3de84045a50464d44e9c39b9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 951.101859] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c319277e-63ff-46fd-a975-49b0644f34a8 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 951.102128] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c319277e-63ff-46fd-a975-49b0644f34a8 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 951.102320] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c319277e-63ff-46fd-a975-49b0644f34a8 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Deleting the datastore file [datastore1] df559cb8-555c-496b-bce3-0981be3ef65f {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 951.102824] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2dac7a2f-796e-4b00-b89c-08d5141dafe6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.109361] env[62385]: DEBUG oslo_vmware.api [None req-c319277e-63ff-46fd-a975-49b0644f34a8 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 951.109361] env[62385]: value = "task-1206172" [ 951.109361] env[62385]: _type = "Task" [ 951.109361] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.117655] env[62385]: DEBUG oslo_vmware.api [None req-c319277e-63ff-46fd-a975-49b0644f34a8 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206172, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.231007] env[62385]: DEBUG nova.network.neutron [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 951.372179] env[62385]: DEBUG nova.network.neutron [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Updating instance_info_cache with network_info: [{"id": "1b04305d-1213-4f55-9b69-6f15de71f576", "address": "fa:16:3e:85:9d:e7", "network": {"id": "afae6b07-d08c-413f-aa73-b545db1bb766", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1654965885-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "93111f5ba17c47de9691f90b99e03aa8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b04305d-12", "ovs_interfaceid": "1b04305d-1213-4f55-9b69-6f15de71f576", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.401214] env[62385]: DEBUG nova.network.neutron [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] [instance: 104dea2c-6977-4683-bc32-05131858f562] Successfully created port: dc5f8505-529f-458b-9f90-58f310ea7ad0 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 951.407270] env[62385]: DEBUG oslo_vmware.api [None req-a86e2e6f-9cfc-4d35-b6f5-b06a789b3209 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206170, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.484388] env[62385]: DEBUG nova.compute.manager [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] [instance: 104dea2c-6977-4683-bc32-05131858f562] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 951.625447] env[62385]: DEBUG oslo_vmware.api [None req-c319277e-63ff-46fd-a975-49b0644f34a8 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206172, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.124681} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.625821] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c319277e-63ff-46fd-a975-49b0644f34a8 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 951.626053] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c319277e-63ff-46fd-a975-49b0644f34a8 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 951.626243] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c319277e-63ff-46fd-a975-49b0644f34a8 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 951.626422] env[62385]: INFO nova.compute.manager [None req-c319277e-63ff-46fd-a975-49b0644f34a8 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Took 1.14 seconds to destroy the instance on the hypervisor. [ 951.626742] env[62385]: DEBUG oslo.service.loopingcall [None req-c319277e-63ff-46fd-a975-49b0644f34a8 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 951.626954] env[62385]: DEBUG nova.compute.manager [-] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 951.627064] env[62385]: DEBUG nova.network.neutron [-] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 951.748952] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fb3ad89-767f-4551-a420-4281c743248f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.756591] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7693773-c798-4bcc-ae3b-9fdb730a93f5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.793273] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-129c3946-1465-41b4-9219-31d639bbffc9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.800787] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f9d409d-7623-4e4e-8fd1-c08bdbc42457 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.814200] env[62385]: DEBUG nova.compute.provider_tree [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 951.876293] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Releasing lock "refresh_cache-5daf0a99-0c2a-40d8-afc7-1998e21e32d8" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.876293] env[62385]: DEBUG nova.compute.manager [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Instance network_info: |[{"id": "1b04305d-1213-4f55-9b69-6f15de71f576", "address": "fa:16:3e:85:9d:e7", "network": {"id": "afae6b07-d08c-413f-aa73-b545db1bb766", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1654965885-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "93111f5ba17c47de9691f90b99e03aa8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b04305d-12", "ovs_interfaceid": "1b04305d-1213-4f55-9b69-6f15de71f576", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 951.876293] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:85:9d:e7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4adc8ed0-d11a-4510-9be0-b27c0da3a903', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1b04305d-1213-4f55-9b69-6f15de71f576', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 951.888110] env[62385]: DEBUG oslo.service.loopingcall [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 951.888711] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 951.888972] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-37bcb65e-e4d3-4dde-aa29-e5479063b3c7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.919901] env[62385]: DEBUG oslo_vmware.api [None req-a86e2e6f-9cfc-4d35-b6f5-b06a789b3209 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206170, 'name': ReconfigVM_Task, 'duration_secs': 0.589624} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.922037] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a86e2e6f-9cfc-4d35-b6f5-b06a789b3209 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Releasing lock "f937ef7b-ab39-4d8a-9577-7faabc652aa5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.923578] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a86e2e6f-9cfc-4d35-b6f5-b06a789b3209 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Reconfigured VM to attach interface {{(pid=62385) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 951.925542] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 951.925542] env[62385]: value = "task-1206173" [ 951.925542] env[62385]: _type = "Task" [ 951.925542] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.934835] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206173, 'name': CreateVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.009061] env[62385]: DEBUG nova.network.neutron [req-272dfbda-e6dc-4b52-aa97-737ba96a3d84 req-b7e1c8db-5960-46a5-8c8a-11c4f637657a service nova] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Updated VIF entry in instance network info cache for port c053ce8b-fc9f-4c48-9073-cdd753630700. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 952.009061] env[62385]: DEBUG nova.network.neutron [req-272dfbda-e6dc-4b52-aa97-737ba96a3d84 req-b7e1c8db-5960-46a5-8c8a-11c4f637657a service nova] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Updating instance_info_cache with network_info: [{"id": "2b05b2b6-d4ce-4eb7-9d65-730d4dbad51c", "address": "fa:16:3e:ec:d7:16", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b05b2b6-d4", "ovs_interfaceid": "2b05b2b6-d4ce-4eb7-9d65-730d4dbad51c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c053ce8b-fc9f-4c48-9073-cdd753630700", "address": "fa:16:3e:9c:f6:87", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc053ce8b-fc", "ovs_interfaceid": "c053ce8b-fc9f-4c48-9073-cdd753630700", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.317662] env[62385]: DEBUG nova.scheduler.client.report [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 952.431016] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a86e2e6f-9cfc-4d35-b6f5-b06a789b3209 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "interface-f937ef7b-ab39-4d8a-9577-7faabc652aa5-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.455s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.435053] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206173, 'name': CreateVM_Task, 'duration_secs': 0.49174} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.435414] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 952.436074] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.436243] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.436557] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 952.436799] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-55b7b88c-82b3-44d2-9394-7cd3bbd02fc2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.441243] env[62385]: DEBUG oslo_vmware.api [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 952.441243] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52736dde-2f7f-bcbb-614c-620c0c672f5e" [ 952.441243] env[62385]: _type = "Task" [ 952.441243] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.448684] env[62385]: DEBUG oslo_vmware.api [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52736dde-2f7f-bcbb-614c-620c0c672f5e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.469108] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9dfd66a2-f29e-47d2-9366-04a5ee79630a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquiring lock "bd7263b2-b996-4794-946b-2c28215574cb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.469525] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9dfd66a2-f29e-47d2-9366-04a5ee79630a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "bd7263b2-b996-4794-946b-2c28215574cb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.469882] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9dfd66a2-f29e-47d2-9366-04a5ee79630a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquiring lock "bd7263b2-b996-4794-946b-2c28215574cb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.470657] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9dfd66a2-f29e-47d2-9366-04a5ee79630a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "bd7263b2-b996-4794-946b-2c28215574cb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.470657] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9dfd66a2-f29e-47d2-9366-04a5ee79630a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "bd7263b2-b996-4794-946b-2c28215574cb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.473937] env[62385]: INFO nova.compute.manager [None req-9dfd66a2-f29e-47d2-9366-04a5ee79630a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Terminating instance [ 952.476583] env[62385]: DEBUG nova.compute.manager [None req-9dfd66a2-f29e-47d2-9366-04a5ee79630a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 952.476897] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-9dfd66a2-f29e-47d2-9366-04a5ee79630a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 952.478158] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21332d06-2795-4e37-b9ce-cfb82b85b4dc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.488770] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-9dfd66a2-f29e-47d2-9366-04a5ee79630a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 952.489135] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cabccc8b-f494-4e11-8243-f4f992df1787 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.495717] env[62385]: DEBUG oslo_vmware.api [None req-9dfd66a2-f29e-47d2-9366-04a5ee79630a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 952.495717] env[62385]: value = "task-1206174" [ 952.495717] env[62385]: _type = "Task" [ 952.495717] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.501684] env[62385]: DEBUG nova.compute.manager [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] [instance: 104dea2c-6977-4683-bc32-05131858f562] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 952.513378] env[62385]: DEBUG nova.network.neutron [-] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.515405] env[62385]: DEBUG oslo_concurrency.lockutils [req-272dfbda-e6dc-4b52-aa97-737ba96a3d84 req-b7e1c8db-5960-46a5-8c8a-11c4f637657a service nova] Releasing lock "refresh_cache-f937ef7b-ab39-4d8a-9577-7faabc652aa5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.515771] env[62385]: DEBUG nova.compute.manager [req-272dfbda-e6dc-4b52-aa97-737ba96a3d84 req-b7e1c8db-5960-46a5-8c8a-11c4f637657a service nova] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Received event network-vif-deleted-65f0aa57-f11b-49fb-8889-1fab7629b3d0 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 952.516329] env[62385]: DEBUG oslo_vmware.api [None req-9dfd66a2-f29e-47d2-9366-04a5ee79630a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206174, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.528592] env[62385]: DEBUG nova.virt.hardware [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 952.528865] env[62385]: DEBUG nova.virt.hardware [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 952.529070] env[62385]: DEBUG nova.virt.hardware [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 952.529302] env[62385]: DEBUG nova.virt.hardware [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 952.529503] env[62385]: DEBUG nova.virt.hardware [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 952.529677] env[62385]: DEBUG nova.virt.hardware [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 952.529921] env[62385]: DEBUG nova.virt.hardware [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 952.530143] env[62385]: DEBUG nova.virt.hardware [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 952.530357] env[62385]: DEBUG nova.virt.hardware [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 952.530557] env[62385]: DEBUG nova.virt.hardware [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 952.530756] env[62385]: DEBUG nova.virt.hardware [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 952.531637] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a884330-c793-46f7-ad27-791dec7a92ca {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.540230] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-262475d5-9231-4416-897b-0217740bcd2c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.692164] env[62385]: DEBUG nova.compute.manager [req-637b6f4b-8152-4be2-bb00-18533940971d req-1b04efe5-c28e-41b9-aaca-4e060afccb3b service nova] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Received event network-changed-1b04305d-1213-4f55-9b69-6f15de71f576 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 952.692164] env[62385]: DEBUG nova.compute.manager [req-637b6f4b-8152-4be2-bb00-18533940971d req-1b04efe5-c28e-41b9-aaca-4e060afccb3b service nova] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Refreshing instance network info cache due to event network-changed-1b04305d-1213-4f55-9b69-6f15de71f576. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 952.692642] env[62385]: DEBUG oslo_concurrency.lockutils [req-637b6f4b-8152-4be2-bb00-18533940971d req-1b04efe5-c28e-41b9-aaca-4e060afccb3b service nova] Acquiring lock "refresh_cache-5daf0a99-0c2a-40d8-afc7-1998e21e32d8" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.692642] env[62385]: DEBUG oslo_concurrency.lockutils [req-637b6f4b-8152-4be2-bb00-18533940971d req-1b04efe5-c28e-41b9-aaca-4e060afccb3b service nova] Acquired lock "refresh_cache-5daf0a99-0c2a-40d8-afc7-1998e21e32d8" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.692858] env[62385]: DEBUG nova.network.neutron [req-637b6f4b-8152-4be2-bb00-18533940971d req-1b04efe5-c28e-41b9-aaca-4e060afccb3b service nova] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Refreshing network info cache for port 1b04305d-1213-4f55-9b69-6f15de71f576 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 952.823073] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.340s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.823536] env[62385]: DEBUG nova.compute.manager [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 952.827045] env[62385]: DEBUG oslo_concurrency.lockutils [None req-84b85202-2a88-4ae6-a721-84e76137e9f4 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.737s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.827160] env[62385]: DEBUG nova.objects.instance [None req-84b85202-2a88-4ae6-a721-84e76137e9f4 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Lazy-loading 'resources' on Instance uuid 75cf7907-d9e3-4f54-90c9-f8d714e1df40 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 952.952891] env[62385]: DEBUG oslo_vmware.api [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52736dde-2f7f-bcbb-614c-620c0c672f5e, 'name': SearchDatastore_Task, 'duration_secs': 0.008925} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.953226] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.953494] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 952.953739] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.953915] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.954222] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 952.954519] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6026443d-5bb8-4a6b-9bd7-83d6c52a2128 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.964014] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 952.964248] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 952.965019] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8c9f2d28-11c8-4fef-b022-6878a017374f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.970356] env[62385]: DEBUG oslo_vmware.api [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 952.970356] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52fbd6cc-d7c7-e65a-b8fd-85d8276d101f" [ 952.970356] env[62385]: _type = "Task" [ 952.970356] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.978346] env[62385]: DEBUG oslo_vmware.api [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52fbd6cc-d7c7-e65a-b8fd-85d8276d101f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.004710] env[62385]: DEBUG oslo_vmware.api [None req-9dfd66a2-f29e-47d2-9366-04a5ee79630a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206174, 'name': PowerOffVM_Task, 'duration_secs': 0.19792} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.004997] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-9dfd66a2-f29e-47d2-9366-04a5ee79630a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 953.005191] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-9dfd66a2-f29e-47d2-9366-04a5ee79630a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 953.005464] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f4225282-9a4f-4a4a-b981-ca538a06bda2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.017732] env[62385]: INFO nova.compute.manager [-] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Took 1.39 seconds to deallocate network for instance. [ 953.079550] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-9dfd66a2-f29e-47d2-9366-04a5ee79630a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 953.079826] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-9dfd66a2-f29e-47d2-9366-04a5ee79630a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Deleting contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 953.080021] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-9dfd66a2-f29e-47d2-9366-04a5ee79630a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Deleting the datastore file [datastore2] bd7263b2-b996-4794-946b-2c28215574cb {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 953.080290] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d5fb9c7d-ef37-467a-9495-06cd5a8d3d0a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.087138] env[62385]: DEBUG oslo_vmware.api [None req-9dfd66a2-f29e-47d2-9366-04a5ee79630a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 953.087138] env[62385]: value = "task-1206176" [ 953.087138] env[62385]: _type = "Task" [ 953.087138] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.095554] env[62385]: DEBUG oslo_vmware.api [None req-9dfd66a2-f29e-47d2-9366-04a5ee79630a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206176, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.287344] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquiring lock "9741e9c8-60d9-41bf-8185-ef3cfcc1c844" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.287753] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "9741e9c8-60d9-41bf-8185-ef3cfcc1c844" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.330340] env[62385]: DEBUG nova.compute.utils [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 953.339457] env[62385]: DEBUG nova.compute.manager [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 953.340325] env[62385]: DEBUG nova.network.neutron [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 953.403033] env[62385]: DEBUG nova.policy [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7e04c6591fe043d686140eb2f5ce0be2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd45b5346eab44707b4d81eaaf12b0935', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 953.486576] env[62385]: DEBUG nova.compute.manager [req-bc0b66f8-1a55-4a1a-8018-232e8c0a5263 req-acca4119-af0c-4550-aa3e-9b6c164a2f80 service nova] [instance: 104dea2c-6977-4683-bc32-05131858f562] Received event network-vif-plugged-dc5f8505-529f-458b-9f90-58f310ea7ad0 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 953.486576] env[62385]: DEBUG oslo_concurrency.lockutils [req-bc0b66f8-1a55-4a1a-8018-232e8c0a5263 req-acca4119-af0c-4550-aa3e-9b6c164a2f80 service nova] Acquiring lock "104dea2c-6977-4683-bc32-05131858f562-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.486576] env[62385]: DEBUG oslo_concurrency.lockutils [req-bc0b66f8-1a55-4a1a-8018-232e8c0a5263 req-acca4119-af0c-4550-aa3e-9b6c164a2f80 service nova] Lock "104dea2c-6977-4683-bc32-05131858f562-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.486576] env[62385]: DEBUG oslo_concurrency.lockutils [req-bc0b66f8-1a55-4a1a-8018-232e8c0a5263 req-acca4119-af0c-4550-aa3e-9b6c164a2f80 service nova] Lock "104dea2c-6977-4683-bc32-05131858f562-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.486576] env[62385]: DEBUG nova.compute.manager [req-bc0b66f8-1a55-4a1a-8018-232e8c0a5263 req-acca4119-af0c-4550-aa3e-9b6c164a2f80 service nova] [instance: 104dea2c-6977-4683-bc32-05131858f562] No waiting events found dispatching network-vif-plugged-dc5f8505-529f-458b-9f90-58f310ea7ad0 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 953.486576] env[62385]: WARNING nova.compute.manager [req-bc0b66f8-1a55-4a1a-8018-232e8c0a5263 req-acca4119-af0c-4550-aa3e-9b6c164a2f80 service nova] [instance: 104dea2c-6977-4683-bc32-05131858f562] Received unexpected event network-vif-plugged-dc5f8505-529f-458b-9f90-58f310ea7ad0 for instance with vm_state building and task_state spawning. [ 953.503329] env[62385]: DEBUG oslo_vmware.api [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52fbd6cc-d7c7-e65a-b8fd-85d8276d101f, 'name': SearchDatastore_Task, 'duration_secs': 0.008901} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.509514] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2e58ce64-0bb7-4444-9ce7-a9c3f7d7c0f1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.517222] env[62385]: DEBUG oslo_vmware.api [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 953.517222] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]522cf47a-fcf0-c03c-d3c3-e8463323c60e" [ 953.517222] env[62385]: _type = "Task" [ 953.517222] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.524928] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c319277e-63ff-46fd-a975-49b0644f34a8 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.528596] env[62385]: DEBUG oslo_vmware.api [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]522cf47a-fcf0-c03c-d3c3-e8463323c60e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.603547] env[62385]: DEBUG oslo_vmware.api [None req-9dfd66a2-f29e-47d2-9366-04a5ee79630a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206176, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138623} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.606070] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-9dfd66a2-f29e-47d2-9366-04a5ee79630a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 953.606276] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-9dfd66a2-f29e-47d2-9366-04a5ee79630a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Deleted contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 953.606541] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-9dfd66a2-f29e-47d2-9366-04a5ee79630a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 953.606629] env[62385]: INFO nova.compute.manager [None req-9dfd66a2-f29e-47d2-9366-04a5ee79630a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Took 1.13 seconds to destroy the instance on the hypervisor. [ 953.606864] env[62385]: DEBUG oslo.service.loopingcall [None req-9dfd66a2-f29e-47d2-9366-04a5ee79630a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 953.607370] env[62385]: DEBUG nova.compute.manager [-] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 953.607458] env[62385]: DEBUG nova.network.neutron [-] [instance: bd7263b2-b996-4794-946b-2c28215574cb] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 953.660274] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f16d324-c9df-416f-b118-d12d48318d1b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.667610] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3fab830-b8e4-4915-a8c9-8978a8e5c12c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.698272] env[62385]: DEBUG nova.network.neutron [req-637b6f4b-8152-4be2-bb00-18533940971d req-1b04efe5-c28e-41b9-aaca-4e060afccb3b service nova] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Updated VIF entry in instance network info cache for port 1b04305d-1213-4f55-9b69-6f15de71f576. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 953.698714] env[62385]: DEBUG nova.network.neutron [req-637b6f4b-8152-4be2-bb00-18533940971d req-1b04efe5-c28e-41b9-aaca-4e060afccb3b service nova] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Updating instance_info_cache with network_info: [{"id": "1b04305d-1213-4f55-9b69-6f15de71f576", "address": "fa:16:3e:85:9d:e7", "network": {"id": "afae6b07-d08c-413f-aa73-b545db1bb766", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1654965885-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "93111f5ba17c47de9691f90b99e03aa8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b04305d-12", "ovs_interfaceid": "1b04305d-1213-4f55-9b69-6f15de71f576", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.700487] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63cb4dcf-bbbf-4e72-9e6e-cfd31dc39723 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.708524] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62f0e257-7468-4482-9757-e49cccea4295 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.724055] env[62385]: DEBUG nova.compute.provider_tree [None req-84b85202-2a88-4ae6-a721-84e76137e9f4 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 953.758219] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Acquiring lock "a1c71d8b-0953-4cc2-961d-0c2b2fc11110" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.758510] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "a1c71d8b-0953-4cc2-961d-0c2b2fc11110" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.790257] env[62385]: DEBUG nova.compute.manager [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 953.829946] env[62385]: DEBUG nova.network.neutron [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Successfully created port: 31a45e90-bf05-4452-8a2c-490aa557ac1b {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 953.839295] env[62385]: DEBUG nova.compute.manager [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 953.971079] env[62385]: DEBUG nova.compute.manager [req-a3513cdc-89b8-4ade-b9ec-437360dac248 req-5dab3a06-97dd-40d7-8846-92a26fe05149 service nova] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Received event network-vif-deleted-fa9f8b71-802a-46a3-a93b-20ee02c6adcb {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 953.971079] env[62385]: INFO nova.compute.manager [req-a3513cdc-89b8-4ade-b9ec-437360dac248 req-5dab3a06-97dd-40d7-8846-92a26fe05149 service nova] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Neutron deleted interface fa9f8b71-802a-46a3-a93b-20ee02c6adcb; detaching it from the instance and deleting it from the info cache [ 953.971079] env[62385]: DEBUG nova.network.neutron [req-a3513cdc-89b8-4ade-b9ec-437360dac248 req-5dab3a06-97dd-40d7-8846-92a26fe05149 service nova] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.027812] env[62385]: DEBUG oslo_vmware.api [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]522cf47a-fcf0-c03c-d3c3-e8463323c60e, 'name': SearchDatastore_Task, 'duration_secs': 0.016546} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.027968] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.028266] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 5daf0a99-0c2a-40d8-afc7-1998e21e32d8/5daf0a99-0c2a-40d8-afc7-1998e21e32d8.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 954.028533] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6075cff6-c65b-4798-95a3-ee4e335b427a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.036235] env[62385]: DEBUG oslo_vmware.api [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 954.036235] env[62385]: value = "task-1206177" [ 954.036235] env[62385]: _type = "Task" [ 954.036235] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.044285] env[62385]: DEBUG oslo_vmware.api [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206177, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.059131] env[62385]: DEBUG nova.network.neutron [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] [instance: 104dea2c-6977-4683-bc32-05131858f562] Successfully updated port: dc5f8505-529f-458b-9f90-58f310ea7ad0 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 954.204048] env[62385]: DEBUG oslo_concurrency.lockutils [req-637b6f4b-8152-4be2-bb00-18533940971d req-1b04efe5-c28e-41b9-aaca-4e060afccb3b service nova] Releasing lock "refresh_cache-5daf0a99-0c2a-40d8-afc7-1998e21e32d8" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.204640] env[62385]: DEBUG nova.compute.manager [req-637b6f4b-8152-4be2-bb00-18533940971d req-1b04efe5-c28e-41b9-aaca-4e060afccb3b service nova] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Received event network-vif-deleted-0533b6c1-8d5b-4822-acd6-a95a651ba50c {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 954.227090] env[62385]: DEBUG nova.scheduler.client.report [None req-84b85202-2a88-4ae6-a721-84e76137e9f4 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 954.262508] env[62385]: DEBUG nova.compute.manager [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 954.315287] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.380767] env[62385]: DEBUG nova.network.neutron [-] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.418490] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a35978b7-46d7-4e06-9dbf-0d023b5874ca tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "interface-f937ef7b-ab39-4d8a-9577-7faabc652aa5-091890dd-d345-4cee-8b38-f7f812092c5f" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.418490] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a35978b7-46d7-4e06-9dbf-0d023b5874ca tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "interface-f937ef7b-ab39-4d8a-9577-7faabc652aa5-091890dd-d345-4cee-8b38-f7f812092c5f" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.418863] env[62385]: DEBUG nova.objects.instance [None req-a35978b7-46d7-4e06-9dbf-0d023b5874ca tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lazy-loading 'flavor' on Instance uuid f937ef7b-ab39-4d8a-9577-7faabc652aa5 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 954.472912] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7d57f138-4ffd-4963-a8d3-a471d08edc31 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.483995] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9669e7c1-d15c-4886-b047-762ef8790084 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.514056] env[62385]: DEBUG nova.compute.manager [req-a3513cdc-89b8-4ade-b9ec-437360dac248 req-5dab3a06-97dd-40d7-8846-92a26fe05149 service nova] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Detach interface failed, port_id=fa9f8b71-802a-46a3-a93b-20ee02c6adcb, reason: Instance bd7263b2-b996-4794-946b-2c28215574cb could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 954.548248] env[62385]: DEBUG oslo_vmware.api [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206177, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.562105] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Acquiring lock "refresh_cache-104dea2c-6977-4683-bc32-05131858f562" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.562344] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Acquired lock "refresh_cache-104dea2c-6977-4683-bc32-05131858f562" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.562642] env[62385]: DEBUG nova.network.neutron [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] [instance: 104dea2c-6977-4683-bc32-05131858f562] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 954.732870] env[62385]: DEBUG oslo_concurrency.lockutils [None req-84b85202-2a88-4ae6-a721-84e76137e9f4 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.906s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.735307] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.324s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.736838] env[62385]: INFO nova.compute.claims [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 954.764427] env[62385]: INFO nova.scheduler.client.report [None req-84b85202-2a88-4ae6-a721-84e76137e9f4 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Deleted allocations for instance 75cf7907-d9e3-4f54-90c9-f8d714e1df40 [ 954.783353] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.849218] env[62385]: DEBUG nova.compute.manager [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 954.879589] env[62385]: DEBUG nova.virt.hardware [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 954.879856] env[62385]: DEBUG nova.virt.hardware [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 954.880027] env[62385]: DEBUG nova.virt.hardware [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 954.880225] env[62385]: DEBUG nova.virt.hardware [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 954.880374] env[62385]: DEBUG nova.virt.hardware [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 954.880524] env[62385]: DEBUG nova.virt.hardware [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 954.880738] env[62385]: DEBUG nova.virt.hardware [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 954.880900] env[62385]: DEBUG nova.virt.hardware [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 954.881084] env[62385]: DEBUG nova.virt.hardware [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 954.881256] env[62385]: DEBUG nova.virt.hardware [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 954.881431] env[62385]: DEBUG nova.virt.hardware [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 954.882612] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-657f4684-def6-40bb-9304-e46058deba76 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.885529] env[62385]: INFO nova.compute.manager [-] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Took 1.28 seconds to deallocate network for instance. [ 954.893105] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61eb242c-8a3a-43a2-89f1-7b318ca1f56b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.048085] env[62385]: DEBUG oslo_vmware.api [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206177, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.590936} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.048085] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 5daf0a99-0c2a-40d8-afc7-1998e21e32d8/5daf0a99-0c2a-40d8-afc7-1998e21e32d8.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 955.048085] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 955.048085] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8d5b91ad-10ae-4689-ae89-7156eea75d4b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.055416] env[62385]: DEBUG nova.objects.instance [None req-a35978b7-46d7-4e06-9dbf-0d023b5874ca tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lazy-loading 'pci_requests' on Instance uuid f937ef7b-ab39-4d8a-9577-7faabc652aa5 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 955.056561] env[62385]: DEBUG oslo_vmware.api [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 955.056561] env[62385]: value = "task-1206178" [ 955.056561] env[62385]: _type = "Task" [ 955.056561] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.068021] env[62385]: DEBUG oslo_vmware.api [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206178, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.132850] env[62385]: DEBUG nova.network.neutron [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] [instance: 104dea2c-6977-4683-bc32-05131858f562] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 955.273735] env[62385]: DEBUG oslo_concurrency.lockutils [None req-84b85202-2a88-4ae6-a721-84e76137e9f4 tempest-ServersTestMultiNic-182692492 tempest-ServersTestMultiNic-182692492-project-member] Lock "75cf7907-d9e3-4f54-90c9-f8d714e1df40" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.471s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.332708] env[62385]: DEBUG nova.network.neutron [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] [instance: 104dea2c-6977-4683-bc32-05131858f562] Updating instance_info_cache with network_info: [{"id": "dc5f8505-529f-458b-9f90-58f310ea7ad0", "address": "fa:16:3e:06:ef:e8", "network": {"id": "8eff1ae2-52e0-452d-8d93-c383c2575716", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-2098602941-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2a9f68fd3de84045a50464d44e9c39b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc5f8505-52", "ovs_interfaceid": "dc5f8505-529f-458b-9f90-58f310ea7ad0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.393599] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9dfd66a2-f29e-47d2-9366-04a5ee79630a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.529180] env[62385]: DEBUG nova.compute.manager [req-60d10772-c2d7-4b0d-b38b-339b5d274de1 req-8e948375-0ac6-4b7d-8206-e98a1ce85132 service nova] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Received event network-vif-plugged-31a45e90-bf05-4452-8a2c-490aa557ac1b {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 955.529180] env[62385]: DEBUG oslo_concurrency.lockutils [req-60d10772-c2d7-4b0d-b38b-339b5d274de1 req-8e948375-0ac6-4b7d-8206-e98a1ce85132 service nova] Acquiring lock "fd0b59c2-e38d-4ecc-b32f-72f07d555ffd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.529389] env[62385]: DEBUG oslo_concurrency.lockutils [req-60d10772-c2d7-4b0d-b38b-339b5d274de1 req-8e948375-0ac6-4b7d-8206-e98a1ce85132 service nova] Lock "fd0b59c2-e38d-4ecc-b32f-72f07d555ffd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.529610] env[62385]: DEBUG oslo_concurrency.lockutils [req-60d10772-c2d7-4b0d-b38b-339b5d274de1 req-8e948375-0ac6-4b7d-8206-e98a1ce85132 service nova] Lock "fd0b59c2-e38d-4ecc-b32f-72f07d555ffd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.530286] env[62385]: DEBUG nova.compute.manager [req-60d10772-c2d7-4b0d-b38b-339b5d274de1 req-8e948375-0ac6-4b7d-8206-e98a1ce85132 service nova] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] No waiting events found dispatching network-vif-plugged-31a45e90-bf05-4452-8a2c-490aa557ac1b {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 955.530286] env[62385]: WARNING nova.compute.manager [req-60d10772-c2d7-4b0d-b38b-339b5d274de1 req-8e948375-0ac6-4b7d-8206-e98a1ce85132 service nova] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Received unexpected event network-vif-plugged-31a45e90-bf05-4452-8a2c-490aa557ac1b for instance with vm_state building and task_state spawning. [ 955.561688] env[62385]: DEBUG nova.objects.base [None req-a35978b7-46d7-4e06-9dbf-0d023b5874ca tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62385) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 955.562102] env[62385]: DEBUG nova.network.neutron [None req-a35978b7-46d7-4e06-9dbf-0d023b5874ca tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 955.571596] env[62385]: DEBUG oslo_vmware.api [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206178, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.100889} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.571854] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 955.573964] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2a9200b-3895-49f2-b521-aac8d1ea9b98 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.595201] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] 5daf0a99-0c2a-40d8-afc7-1998e21e32d8/5daf0a99-0c2a-40d8-afc7-1998e21e32d8.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 955.596533] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-886c32b3-ade6-4f01-a5ec-37cdadacf149 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.619867] env[62385]: DEBUG oslo_vmware.api [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 955.619867] env[62385]: value = "task-1206179" [ 955.619867] env[62385]: _type = "Task" [ 955.619867] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.628223] env[62385]: DEBUG oslo_vmware.api [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206179, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.667547] env[62385]: DEBUG nova.policy [None req-a35978b7-46d7-4e06-9dbf-0d023b5874ca tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5040b20c848c4d28a198f1773a3ff21e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '012426f098ce40c3aaa00f628fe9cebb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 955.835966] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Releasing lock "refresh_cache-104dea2c-6977-4683-bc32-05131858f562" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.836350] env[62385]: DEBUG nova.compute.manager [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] [instance: 104dea2c-6977-4683-bc32-05131858f562] Instance network_info: |[{"id": "dc5f8505-529f-458b-9f90-58f310ea7ad0", "address": "fa:16:3e:06:ef:e8", "network": {"id": "8eff1ae2-52e0-452d-8d93-c383c2575716", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-2098602941-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2a9f68fd3de84045a50464d44e9c39b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc5f8505-52", "ovs_interfaceid": "dc5f8505-529f-458b-9f90-58f310ea7ad0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 955.836991] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] [instance: 104dea2c-6977-4683-bc32-05131858f562] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:06:ef:e8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f77ff7a1-209c-4f3f-b2a0-fd817741e739', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dc5f8505-529f-458b-9f90-58f310ea7ad0', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 955.845324] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Creating folder: Project (2a9f68fd3de84045a50464d44e9c39b9). Parent ref: group-v261107. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 955.847912] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aa805e8d-a112-44e1-a3a7-dd57d580f966 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.860527] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Created folder: Project (2a9f68fd3de84045a50464d44e9c39b9) in parent group-v261107. [ 955.860527] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Creating folder: Instances. Parent ref: group-v261265. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 955.860527] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4fb3e4b7-cbc3-480a-b8d0-87674af48747 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.872463] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Created folder: Instances in parent group-v261265. [ 955.872720] env[62385]: DEBUG oslo.service.loopingcall [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 955.872909] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 104dea2c-6977-4683-bc32-05131858f562] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 955.873129] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f11c9b63-b7ae-48d7-90a4-f6638355bedc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.896115] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 955.896115] env[62385]: value = "task-1206182" [ 955.896115] env[62385]: _type = "Task" [ 955.896115] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.906945] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206182, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.932171] env[62385]: DEBUG nova.network.neutron [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Successfully updated port: 31a45e90-bf05-4452-8a2c-490aa557ac1b {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 956.029413] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-374e9220-4447-44b0-a482-f0a57762c175 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.037171] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eabdabf-f7ff-4f9f-82f1-b91c7c214393 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.071922] env[62385]: DEBUG nova.compute.manager [req-949d4f2b-4336-4652-86ca-bdf88a6e572f req-f18c8953-3bbb-4295-bd8f-3a84c4b48375 service nova] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Received event network-changed-31a45e90-bf05-4452-8a2c-490aa557ac1b {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 956.072151] env[62385]: DEBUG nova.compute.manager [req-949d4f2b-4336-4652-86ca-bdf88a6e572f req-f18c8953-3bbb-4295-bd8f-3a84c4b48375 service nova] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Refreshing instance network info cache due to event network-changed-31a45e90-bf05-4452-8a2c-490aa557ac1b. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 956.072374] env[62385]: DEBUG oslo_concurrency.lockutils [req-949d4f2b-4336-4652-86ca-bdf88a6e572f req-f18c8953-3bbb-4295-bd8f-3a84c4b48375 service nova] Acquiring lock "refresh_cache-fd0b59c2-e38d-4ecc-b32f-72f07d555ffd" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.072522] env[62385]: DEBUG oslo_concurrency.lockutils [req-949d4f2b-4336-4652-86ca-bdf88a6e572f req-f18c8953-3bbb-4295-bd8f-3a84c4b48375 service nova] Acquired lock "refresh_cache-fd0b59c2-e38d-4ecc-b32f-72f07d555ffd" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.072685] env[62385]: DEBUG nova.network.neutron [req-949d4f2b-4336-4652-86ca-bdf88a6e572f req-f18c8953-3bbb-4295-bd8f-3a84c4b48375 service nova] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Refreshing network info cache for port 31a45e90-bf05-4452-8a2c-490aa557ac1b {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 956.074202] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e0add6b-1f10-464d-b18a-99516a088aa3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.083869] env[62385]: DEBUG nova.compute.manager [req-10e91ab4-867c-4b07-89af-7ab8b0d537cc req-ca390604-e99f-4771-a3a4-c5eb4f6074cb service nova] [instance: 104dea2c-6977-4683-bc32-05131858f562] Received event network-changed-dc5f8505-529f-458b-9f90-58f310ea7ad0 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 956.084083] env[62385]: DEBUG nova.compute.manager [req-10e91ab4-867c-4b07-89af-7ab8b0d537cc req-ca390604-e99f-4771-a3a4-c5eb4f6074cb service nova] [instance: 104dea2c-6977-4683-bc32-05131858f562] Refreshing instance network info cache due to event network-changed-dc5f8505-529f-458b-9f90-58f310ea7ad0. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 956.084304] env[62385]: DEBUG oslo_concurrency.lockutils [req-10e91ab4-867c-4b07-89af-7ab8b0d537cc req-ca390604-e99f-4771-a3a4-c5eb4f6074cb service nova] Acquiring lock "refresh_cache-104dea2c-6977-4683-bc32-05131858f562" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.084450] env[62385]: DEBUG oslo_concurrency.lockutils [req-10e91ab4-867c-4b07-89af-7ab8b0d537cc req-ca390604-e99f-4771-a3a4-c5eb4f6074cb service nova] Acquired lock "refresh_cache-104dea2c-6977-4683-bc32-05131858f562" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.084613] env[62385]: DEBUG nova.network.neutron [req-10e91ab4-867c-4b07-89af-7ab8b0d537cc req-ca390604-e99f-4771-a3a4-c5eb4f6074cb service nova] [instance: 104dea2c-6977-4683-bc32-05131858f562] Refreshing network info cache for port dc5f8505-529f-458b-9f90-58f310ea7ad0 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 956.086569] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df53af53-01c1-4385-8430-986b779e3f24 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.102894] env[62385]: DEBUG nova.compute.provider_tree [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 956.129913] env[62385]: DEBUG oslo_vmware.api [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206179, 'name': ReconfigVM_Task, 'duration_secs': 0.340324} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.130218] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Reconfigured VM instance instance-0000005d to attach disk [datastore1] 5daf0a99-0c2a-40d8-afc7-1998e21e32d8/5daf0a99-0c2a-40d8-afc7-1998e21e32d8.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 956.130859] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-059e80c4-6590-4845-b825-4737ca864e27 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.137074] env[62385]: DEBUG oslo_vmware.api [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 956.137074] env[62385]: value = "task-1206183" [ 956.137074] env[62385]: _type = "Task" [ 956.137074] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.145044] env[62385]: DEBUG oslo_vmware.api [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206183, 'name': Rename_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.407789] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206182, 'name': CreateVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.437599] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "refresh_cache-fd0b59c2-e38d-4ecc-b32f-72f07d555ffd" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.606057] env[62385]: DEBUG nova.scheduler.client.report [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 956.615705] env[62385]: DEBUG nova.network.neutron [req-949d4f2b-4336-4652-86ca-bdf88a6e572f req-f18c8953-3bbb-4295-bd8f-3a84c4b48375 service nova] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 956.649252] env[62385]: DEBUG oslo_vmware.api [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206183, 'name': Rename_Task, 'duration_secs': 0.142891} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.649252] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 956.649614] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-86b3b49f-cad3-404a-978d-4a865a13c4c4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.656850] env[62385]: DEBUG oslo_vmware.api [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 956.656850] env[62385]: value = "task-1206184" [ 956.656850] env[62385]: _type = "Task" [ 956.656850] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.666740] env[62385]: DEBUG oslo_vmware.api [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206184, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.756307] env[62385]: DEBUG nova.network.neutron [req-949d4f2b-4336-4652-86ca-bdf88a6e572f req-f18c8953-3bbb-4295-bd8f-3a84c4b48375 service nova] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.818545] env[62385]: DEBUG nova.network.neutron [req-10e91ab4-867c-4b07-89af-7ab8b0d537cc req-ca390604-e99f-4771-a3a4-c5eb4f6074cb service nova] [instance: 104dea2c-6977-4683-bc32-05131858f562] Updated VIF entry in instance network info cache for port dc5f8505-529f-458b-9f90-58f310ea7ad0. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 956.818931] env[62385]: DEBUG nova.network.neutron [req-10e91ab4-867c-4b07-89af-7ab8b0d537cc req-ca390604-e99f-4771-a3a4-c5eb4f6074cb service nova] [instance: 104dea2c-6977-4683-bc32-05131858f562] Updating instance_info_cache with network_info: [{"id": "dc5f8505-529f-458b-9f90-58f310ea7ad0", "address": "fa:16:3e:06:ef:e8", "network": {"id": "8eff1ae2-52e0-452d-8d93-c383c2575716", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-2098602941-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2a9f68fd3de84045a50464d44e9c39b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc5f8505-52", "ovs_interfaceid": "dc5f8505-529f-458b-9f90-58f310ea7ad0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.908364] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206182, 'name': CreateVM_Task, 'duration_secs': 0.648584} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.908896] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 104dea2c-6977-4683-bc32-05131858f562] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 956.909225] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.909394] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.909752] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 956.910309] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11b33b3c-d111-457f-824e-dccf71655b0d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.915282] env[62385]: DEBUG oslo_vmware.api [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Waiting for the task: (returnval){ [ 956.915282] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]522ae1cc-af7f-fe97-b83d-0b9de2854536" [ 956.915282] env[62385]: _type = "Task" [ 956.915282] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.922755] env[62385]: DEBUG oslo_vmware.api [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]522ae1cc-af7f-fe97-b83d-0b9de2854536, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.111801] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.376s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.112500] env[62385]: DEBUG nova.compute.manager [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 957.119030] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e6dd994d-385e-4b5b-a2e3-085be93a7d91 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.204s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.119030] env[62385]: DEBUG nova.objects.instance [None req-e6dd994d-385e-4b5b-a2e3-085be93a7d91 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lazy-loading 'resources' on Instance uuid a58cb15f-e4df-4b1e-b09c-2f61dce7200f {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 957.168036] env[62385]: DEBUG oslo_vmware.api [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206184, 'name': PowerOnVM_Task, 'duration_secs': 0.483968} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.168356] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 957.168550] env[62385]: INFO nova.compute.manager [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Took 6.70 seconds to spawn the instance on the hypervisor. [ 957.168734] env[62385]: DEBUG nova.compute.manager [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 957.169557] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f97f0c5c-0d3c-43c0-9eb9-9e06f0d37413 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.261430] env[62385]: DEBUG oslo_concurrency.lockutils [req-949d4f2b-4336-4652-86ca-bdf88a6e572f req-f18c8953-3bbb-4295-bd8f-3a84c4b48375 service nova] Releasing lock "refresh_cache-fd0b59c2-e38d-4ecc-b32f-72f07d555ffd" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.261942] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquired lock "refresh_cache-fd0b59c2-e38d-4ecc-b32f-72f07d555ffd" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.262459] env[62385]: DEBUG nova.network.neutron [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 957.322527] env[62385]: DEBUG oslo_concurrency.lockutils [req-10e91ab4-867c-4b07-89af-7ab8b0d537cc req-ca390604-e99f-4771-a3a4-c5eb4f6074cb service nova] Releasing lock "refresh_cache-104dea2c-6977-4683-bc32-05131858f562" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.345456] env[62385]: DEBUG nova.network.neutron [None req-a35978b7-46d7-4e06-9dbf-0d023b5874ca tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Successfully updated port: 091890dd-d345-4cee-8b38-f7f812092c5f {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 957.427875] env[62385]: DEBUG oslo_vmware.api [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]522ae1cc-af7f-fe97-b83d-0b9de2854536, 'name': SearchDatastore_Task, 'duration_secs': 0.007991} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.428317] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.428689] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] [instance: 104dea2c-6977-4683-bc32-05131858f562] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 957.428878] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.429030] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.429223] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 957.429529] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6072e571-9ce8-43eb-9587-7848ca34d431 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.437819] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 957.438044] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 957.438787] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a3b41307-e6e0-47ed-be93-458e5a12e17d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.445973] env[62385]: DEBUG oslo_vmware.api [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Waiting for the task: (returnval){ [ 957.445973] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5207f32d-c0d7-5fa6-2520-3aa1b0b4f508" [ 957.445973] env[62385]: _type = "Task" [ 957.445973] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.453673] env[62385]: DEBUG oslo_vmware.api [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5207f32d-c0d7-5fa6-2520-3aa1b0b4f508, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.621482] env[62385]: DEBUG nova.compute.utils [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 957.624316] env[62385]: DEBUG nova.compute.manager [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 957.624661] env[62385]: DEBUG nova.network.neutron [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 957.692523] env[62385]: INFO nova.compute.manager [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Took 25.02 seconds to build instance. [ 957.709708] env[62385]: DEBUG nova.policy [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4ac552dc01fc4de2b97e36b41dae725a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '105150c170d74e268e82ab894e9bf0d5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 957.827618] env[62385]: DEBUG nova.network.neutron [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 957.849279] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a35978b7-46d7-4e06-9dbf-0d023b5874ca tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "refresh_cache-f937ef7b-ab39-4d8a-9577-7faabc652aa5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.849462] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a35978b7-46d7-4e06-9dbf-0d023b5874ca tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquired lock "refresh_cache-f937ef7b-ab39-4d8a-9577-7faabc652aa5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.849668] env[62385]: DEBUG nova.network.neutron [None req-a35978b7-46d7-4e06-9dbf-0d023b5874ca tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 957.944687] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5e0bb34-817e-433f-aaa4-85a597696f32 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.956905] env[62385]: DEBUG oslo_vmware.api [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5207f32d-c0d7-5fa6-2520-3aa1b0b4f508, 'name': SearchDatastore_Task, 'duration_secs': 0.008962} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.960834] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-386d60ba-2207-414c-8de4-f2792716c47f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.965068] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d9cfe3d-3142-463c-b9ab-eef030c72cfe {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.997663] env[62385]: DEBUG oslo_vmware.api [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Waiting for the task: (returnval){ [ 957.997663] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52cc8f2c-2ae0-a499-9fcf-1c3b2f46c4d5" [ 957.997663] env[62385]: _type = "Task" [ 957.997663] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.001898] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-858556dd-3c8e-459e-a9de-62ff6a24dc8d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.017780] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41645878-d94f-49e5-8a2e-a52c3363c7ba {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.021773] env[62385]: DEBUG oslo_vmware.api [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52cc8f2c-2ae0-a499-9fcf-1c3b2f46c4d5, 'name': SearchDatastore_Task, 'duration_secs': 0.010478} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.022868] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.023180] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 104dea2c-6977-4683-bc32-05131858f562/104dea2c-6977-4683-bc32-05131858f562.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 958.023817] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1336d534-abac-42b0-ae27-bee3b78a3156 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.033735] env[62385]: DEBUG nova.compute.provider_tree [None req-e6dd994d-385e-4b5b-a2e3-085be93a7d91 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 958.039258] env[62385]: DEBUG oslo_vmware.api [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Waiting for the task: (returnval){ [ 958.039258] env[62385]: value = "task-1206185" [ 958.039258] env[62385]: _type = "Task" [ 958.039258] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.046691] env[62385]: DEBUG oslo_vmware.api [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Task: {'id': task-1206185, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.128641] env[62385]: DEBUG nova.compute.manager [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 958.195125] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2b5277e6-73cb-4f09-9700-dd37efa971d0 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "5daf0a99-0c2a-40d8-afc7-1998e21e32d8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.534s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.269980] env[62385]: DEBUG nova.network.neutron [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Updating instance_info_cache with network_info: [{"id": "31a45e90-bf05-4452-8a2c-490aa557ac1b", "address": "fa:16:3e:7d:b9:30", "network": {"id": "06017161-984c-4657-9639-4e6bdb6d4c09", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1072882119-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d45b5346eab44707b4d81eaaf12b0935", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39cd75b0-9ec7-48ed-b57f-34da0c573a60", "external-id": "nsx-vlan-transportzone-751", "segmentation_id": 751, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap31a45e90-bf", "ovs_interfaceid": "31a45e90-bf05-4452-8a2c-490aa557ac1b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.317877] env[62385]: DEBUG nova.compute.manager [req-12e381ce-9760-4b6c-acea-8bcedbf41da3 req-bedf00cd-ec98-44ed-854e-918e4f90d728 service nova] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Received event network-vif-plugged-091890dd-d345-4cee-8b38-f7f812092c5f {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 958.317877] env[62385]: DEBUG oslo_concurrency.lockutils [req-12e381ce-9760-4b6c-acea-8bcedbf41da3 req-bedf00cd-ec98-44ed-854e-918e4f90d728 service nova] Acquiring lock "f937ef7b-ab39-4d8a-9577-7faabc652aa5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.318185] env[62385]: DEBUG oslo_concurrency.lockutils [req-12e381ce-9760-4b6c-acea-8bcedbf41da3 req-bedf00cd-ec98-44ed-854e-918e4f90d728 service nova] Lock "f937ef7b-ab39-4d8a-9577-7faabc652aa5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.318316] env[62385]: DEBUG oslo_concurrency.lockutils [req-12e381ce-9760-4b6c-acea-8bcedbf41da3 req-bedf00cd-ec98-44ed-854e-918e4f90d728 service nova] Lock "f937ef7b-ab39-4d8a-9577-7faabc652aa5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.318779] env[62385]: DEBUG nova.compute.manager [req-12e381ce-9760-4b6c-acea-8bcedbf41da3 req-bedf00cd-ec98-44ed-854e-918e4f90d728 service nova] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] No waiting events found dispatching network-vif-plugged-091890dd-d345-4cee-8b38-f7f812092c5f {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 958.318779] env[62385]: WARNING nova.compute.manager [req-12e381ce-9760-4b6c-acea-8bcedbf41da3 req-bedf00cd-ec98-44ed-854e-918e4f90d728 service nova] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Received unexpected event network-vif-plugged-091890dd-d345-4cee-8b38-f7f812092c5f for instance with vm_state active and task_state None. [ 958.318870] env[62385]: DEBUG nova.compute.manager [req-12e381ce-9760-4b6c-acea-8bcedbf41da3 req-bedf00cd-ec98-44ed-854e-918e4f90d728 service nova] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Received event network-changed-091890dd-d345-4cee-8b38-f7f812092c5f {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 958.318960] env[62385]: DEBUG nova.compute.manager [req-12e381ce-9760-4b6c-acea-8bcedbf41da3 req-bedf00cd-ec98-44ed-854e-918e4f90d728 service nova] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Refreshing instance network info cache due to event network-changed-091890dd-d345-4cee-8b38-f7f812092c5f. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 958.319349] env[62385]: DEBUG oslo_concurrency.lockutils [req-12e381ce-9760-4b6c-acea-8bcedbf41da3 req-bedf00cd-ec98-44ed-854e-918e4f90d728 service nova] Acquiring lock "refresh_cache-f937ef7b-ab39-4d8a-9577-7faabc652aa5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.421475] env[62385]: WARNING nova.network.neutron [None req-a35978b7-46d7-4e06-9dbf-0d023b5874ca tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] daad7070-dc06-4580-9f4a-ac8aaa546955 already exists in list: networks containing: ['daad7070-dc06-4580-9f4a-ac8aaa546955']. ignoring it [ 958.421702] env[62385]: WARNING nova.network.neutron [None req-a35978b7-46d7-4e06-9dbf-0d023b5874ca tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] daad7070-dc06-4580-9f4a-ac8aaa546955 already exists in list: networks containing: ['daad7070-dc06-4580-9f4a-ac8aaa546955']. ignoring it [ 958.536866] env[62385]: DEBUG nova.scheduler.client.report [None req-e6dd994d-385e-4b5b-a2e3-085be93a7d91 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 958.549338] env[62385]: DEBUG oslo_vmware.api [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Task: {'id': task-1206185, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.464428} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.549657] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 104dea2c-6977-4683-bc32-05131858f562/104dea2c-6977-4683-bc32-05131858f562.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 958.549885] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] [instance: 104dea2c-6977-4683-bc32-05131858f562] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 958.550164] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-531a9ffd-c93f-4865-9a29-cd8341e91269 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.557136] env[62385]: DEBUG oslo_vmware.api [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Waiting for the task: (returnval){ [ 958.557136] env[62385]: value = "task-1206186" [ 958.557136] env[62385]: _type = "Task" [ 958.557136] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.570554] env[62385]: DEBUG oslo_vmware.api [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Task: {'id': task-1206186, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.772326] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Releasing lock "refresh_cache-fd0b59c2-e38d-4ecc-b32f-72f07d555ffd" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.772644] env[62385]: DEBUG nova.compute.manager [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Instance network_info: |[{"id": "31a45e90-bf05-4452-8a2c-490aa557ac1b", "address": "fa:16:3e:7d:b9:30", "network": {"id": "06017161-984c-4657-9639-4e6bdb6d4c09", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1072882119-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d45b5346eab44707b4d81eaaf12b0935", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39cd75b0-9ec7-48ed-b57f-34da0c573a60", "external-id": "nsx-vlan-transportzone-751", "segmentation_id": 751, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap31a45e90-bf", "ovs_interfaceid": "31a45e90-bf05-4452-8a2c-490aa557ac1b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 958.773093] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7d:b9:30', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '39cd75b0-9ec7-48ed-b57f-34da0c573a60', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '31a45e90-bf05-4452-8a2c-490aa557ac1b', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 958.780699] env[62385]: DEBUG oslo.service.loopingcall [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 958.781296] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 958.781544] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1b74b64c-bb4a-41f6-9c3f-9bc2ede0282a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.801143] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 958.801143] env[62385]: value = "task-1206187" [ 958.801143] env[62385]: _type = "Task" [ 958.801143] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.808888] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206187, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.878720] env[62385]: DEBUG nova.network.neutron [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Successfully created port: a9513285-4023-4f3e-b90d-3eb507955f13 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 959.045349] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e6dd994d-385e-4b5b-a2e3-085be93a7d91 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.929s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.049266] env[62385]: DEBUG oslo_concurrency.lockutils [None req-782e1513-c3dc-4b59-8b0f-73e617258048 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.574s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.050328] env[62385]: DEBUG nova.objects.instance [None req-782e1513-c3dc-4b59-8b0f-73e617258048 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lazy-loading 'resources' on Instance uuid f1d33401-35f5-4d79-abb4-26dc6faa784e {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 959.069424] env[62385]: DEBUG oslo_vmware.api [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Task: {'id': task-1206186, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083649} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.069712] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] [instance: 104dea2c-6977-4683-bc32-05131858f562] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 959.070616] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19673d27-9fd1-4de2-99a5-f5dc50bbbe0f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.093648] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] [instance: 104dea2c-6977-4683-bc32-05131858f562] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] 104dea2c-6977-4683-bc32-05131858f562/104dea2c-6977-4683-bc32-05131858f562.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 959.094757] env[62385]: INFO nova.scheduler.client.report [None req-e6dd994d-385e-4b5b-a2e3-085be93a7d91 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Deleted allocations for instance a58cb15f-e4df-4b1e-b09c-2f61dce7200f [ 959.095771] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b5984dcb-8c7d-4808-b01a-759c1c150260 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.122196] env[62385]: DEBUG oslo_vmware.api [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Waiting for the task: (returnval){ [ 959.122196] env[62385]: value = "task-1206188" [ 959.122196] env[62385]: _type = "Task" [ 959.122196] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.128574] env[62385]: DEBUG oslo_vmware.api [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Task: {'id': task-1206188, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.143611] env[62385]: DEBUG nova.compute.manager [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 959.176662] env[62385]: DEBUG nova.virt.hardware [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 959.176940] env[62385]: DEBUG nova.virt.hardware [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 959.177130] env[62385]: DEBUG nova.virt.hardware [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 959.177375] env[62385]: DEBUG nova.virt.hardware [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 959.177486] env[62385]: DEBUG nova.virt.hardware [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 959.177638] env[62385]: DEBUG nova.virt.hardware [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 959.177856] env[62385]: DEBUG nova.virt.hardware [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 959.178375] env[62385]: DEBUG nova.virt.hardware [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 959.178375] env[62385]: DEBUG nova.virt.hardware [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 959.178375] env[62385]: DEBUG nova.virt.hardware [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 959.178618] env[62385]: DEBUG nova.virt.hardware [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 959.179579] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83b7a9f7-af13-4bd8-9c45-32e6aa4b8018 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.192120] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eb732d6-e23e-41d8-b786-638d318f3451 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.311033] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206187, 'name': CreateVM_Task, 'duration_secs': 0.32452} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.311219] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 959.311890] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.312075] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.312395] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 959.312641] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-266eb05d-d2c1-41d7-a7e1-ae7381625289 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.316800] env[62385]: DEBUG oslo_vmware.api [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 959.316800] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]529fee66-34eb-e934-0844-92e842cb17c2" [ 959.316800] env[62385]: _type = "Task" [ 959.316800] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.324363] env[62385]: DEBUG oslo_vmware.api [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]529fee66-34eb-e934-0844-92e842cb17c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.548767] env[62385]: DEBUG nova.network.neutron [None req-a35978b7-46d7-4e06-9dbf-0d023b5874ca tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Updating instance_info_cache with network_info: [{"id": "2b05b2b6-d4ce-4eb7-9d65-730d4dbad51c", "address": "fa:16:3e:ec:d7:16", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b05b2b6-d4", "ovs_interfaceid": "2b05b2b6-d4ce-4eb7-9d65-730d4dbad51c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c053ce8b-fc9f-4c48-9073-cdd753630700", "address": "fa:16:3e:9c:f6:87", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc053ce8b-fc", "ovs_interfaceid": "c053ce8b-fc9f-4c48-9073-cdd753630700", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "091890dd-d345-4cee-8b38-f7f812092c5f", "address": "fa:16:3e:b3:d9:f5", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap091890dd-d3", "ovs_interfaceid": "091890dd-d345-4cee-8b38-f7f812092c5f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.619908] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e6dd994d-385e-4b5b-a2e3-085be93a7d91 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "a58cb15f-e4df-4b1e-b09c-2f61dce7200f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.737s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.629955] env[62385]: DEBUG oslo_vmware.api [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Task: {'id': task-1206188, 'name': ReconfigVM_Task, 'duration_secs': 0.259885} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.632732] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] [instance: 104dea2c-6977-4683-bc32-05131858f562] Reconfigured VM instance instance-0000005e to attach disk [datastore1] 104dea2c-6977-4683-bc32-05131858f562/104dea2c-6977-4683-bc32-05131858f562.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 959.633695] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c13a1f13-104e-4778-b4c5-f389a0422e8e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.640815] env[62385]: DEBUG oslo_vmware.api [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Waiting for the task: (returnval){ [ 959.640815] env[62385]: value = "task-1206189" [ 959.640815] env[62385]: _type = "Task" [ 959.640815] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.652409] env[62385]: DEBUG oslo_vmware.api [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Task: {'id': task-1206189, 'name': Rename_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.771986] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfbfa915-7391-4c5f-a25f-4afbcb74aed1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.779748] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8554188-0dfa-4fbd-93ae-c452f61b5996 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.810898] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0044d45-0199-4ce3-afbe-426074b7fc24 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.818085] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b0ce39f-0ab9-4c64-bcc2-c33e96e778a3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.834431] env[62385]: DEBUG nova.compute.provider_tree [None req-782e1513-c3dc-4b59-8b0f-73e617258048 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 959.839117] env[62385]: DEBUG oslo_vmware.api [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]529fee66-34eb-e934-0844-92e842cb17c2, 'name': SearchDatastore_Task, 'duration_secs': 0.016783} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.839665] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.839900] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 959.840186] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.840358] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.840548] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 959.840840] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1b9ebe18-abb5-484c-8ae1-8124a97d8b80 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.849310] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 959.849521] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 959.850287] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ae21b6a-a708-4b91-8c9b-8d1327a234e6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.857277] env[62385]: DEBUG oslo_vmware.api [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 959.857277] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52d93e21-b6df-8600-9d5b-8a1fda0b7889" [ 959.857277] env[62385]: _type = "Task" [ 959.857277] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.864062] env[62385]: DEBUG oslo_vmware.api [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52d93e21-b6df-8600-9d5b-8a1fda0b7889, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.052355] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a35978b7-46d7-4e06-9dbf-0d023b5874ca tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Releasing lock "refresh_cache-f937ef7b-ab39-4d8a-9577-7faabc652aa5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.053163] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a35978b7-46d7-4e06-9dbf-0d023b5874ca tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "f937ef7b-ab39-4d8a-9577-7faabc652aa5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.053260] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a35978b7-46d7-4e06-9dbf-0d023b5874ca tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquired lock "f937ef7b-ab39-4d8a-9577-7faabc652aa5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.053529] env[62385]: DEBUG oslo_concurrency.lockutils [req-12e381ce-9760-4b6c-acea-8bcedbf41da3 req-bedf00cd-ec98-44ed-854e-918e4f90d728 service nova] Acquired lock "refresh_cache-f937ef7b-ab39-4d8a-9577-7faabc652aa5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.053717] env[62385]: DEBUG nova.network.neutron [req-12e381ce-9760-4b6c-acea-8bcedbf41da3 req-bedf00cd-ec98-44ed-854e-918e4f90d728 service nova] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Refreshing network info cache for port 091890dd-d345-4cee-8b38-f7f812092c5f {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 960.056037] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfa18e4e-b04f-4350-8a1a-3c71bc543077 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.073918] env[62385]: DEBUG nova.virt.hardware [None req-a35978b7-46d7-4e06-9dbf-0d023b5874ca tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 960.073918] env[62385]: DEBUG nova.virt.hardware [None req-a35978b7-46d7-4e06-9dbf-0d023b5874ca tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 960.073918] env[62385]: DEBUG nova.virt.hardware [None req-a35978b7-46d7-4e06-9dbf-0d023b5874ca tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 960.074184] env[62385]: DEBUG nova.virt.hardware [None req-a35978b7-46d7-4e06-9dbf-0d023b5874ca tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 960.074184] env[62385]: DEBUG nova.virt.hardware [None req-a35978b7-46d7-4e06-9dbf-0d023b5874ca tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 960.074550] env[62385]: DEBUG nova.virt.hardware [None req-a35978b7-46d7-4e06-9dbf-0d023b5874ca tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 960.074550] env[62385]: DEBUG nova.virt.hardware [None req-a35978b7-46d7-4e06-9dbf-0d023b5874ca tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 960.074680] env[62385]: DEBUG nova.virt.hardware [None req-a35978b7-46d7-4e06-9dbf-0d023b5874ca tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 960.074926] env[62385]: DEBUG nova.virt.hardware [None req-a35978b7-46d7-4e06-9dbf-0d023b5874ca tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 960.075111] env[62385]: DEBUG nova.virt.hardware [None req-a35978b7-46d7-4e06-9dbf-0d023b5874ca tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 960.075290] env[62385]: DEBUG nova.virt.hardware [None req-a35978b7-46d7-4e06-9dbf-0d023b5874ca tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 960.081771] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a35978b7-46d7-4e06-9dbf-0d023b5874ca tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Reconfiguring VM to attach interface {{(pid=62385) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 960.082844] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d0d1a2a3-183a-4611-958b-4a6e2536b236 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.100388] env[62385]: DEBUG oslo_vmware.api [None req-a35978b7-46d7-4e06-9dbf-0d023b5874ca tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 960.100388] env[62385]: value = "task-1206190" [ 960.100388] env[62385]: _type = "Task" [ 960.100388] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.110451] env[62385]: DEBUG oslo_vmware.api [None req-a35978b7-46d7-4e06-9dbf-0d023b5874ca tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206190, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.151796] env[62385]: DEBUG oslo_vmware.api [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Task: {'id': task-1206189, 'name': Rename_Task, 'duration_secs': 0.137573} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.152098] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] [instance: 104dea2c-6977-4683-bc32-05131858f562] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 960.152359] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2baf94d5-2939-469d-b726-a584b03bbcb5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.159797] env[62385]: DEBUG oslo_vmware.api [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Waiting for the task: (returnval){ [ 960.159797] env[62385]: value = "task-1206191" [ 960.159797] env[62385]: _type = "Task" [ 960.159797] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.168966] env[62385]: DEBUG oslo_vmware.api [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Task: {'id': task-1206191, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.341059] env[62385]: DEBUG nova.scheduler.client.report [None req-782e1513-c3dc-4b59-8b0f-73e617258048 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 960.348671] env[62385]: DEBUG nova.compute.manager [req-f57c1736-6b38-4e21-8135-c2ef0cf0f13f req-b5349e81-495e-401f-9f34-c1f01d85cf6a service nova] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Received event network-changed-1b04305d-1213-4f55-9b69-6f15de71f576 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 960.348671] env[62385]: DEBUG nova.compute.manager [req-f57c1736-6b38-4e21-8135-c2ef0cf0f13f req-b5349e81-495e-401f-9f34-c1f01d85cf6a service nova] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Refreshing instance network info cache due to event network-changed-1b04305d-1213-4f55-9b69-6f15de71f576. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 960.348671] env[62385]: DEBUG oslo_concurrency.lockutils [req-f57c1736-6b38-4e21-8135-c2ef0cf0f13f req-b5349e81-495e-401f-9f34-c1f01d85cf6a service nova] Acquiring lock "refresh_cache-5daf0a99-0c2a-40d8-afc7-1998e21e32d8" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.348671] env[62385]: DEBUG oslo_concurrency.lockutils [req-f57c1736-6b38-4e21-8135-c2ef0cf0f13f req-b5349e81-495e-401f-9f34-c1f01d85cf6a service nova] Acquired lock "refresh_cache-5daf0a99-0c2a-40d8-afc7-1998e21e32d8" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.348671] env[62385]: DEBUG nova.network.neutron [req-f57c1736-6b38-4e21-8135-c2ef0cf0f13f req-b5349e81-495e-401f-9f34-c1f01d85cf6a service nova] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Refreshing network info cache for port 1b04305d-1213-4f55-9b69-6f15de71f576 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 960.367033] env[62385]: DEBUG oslo_vmware.api [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52d93e21-b6df-8600-9d5b-8a1fda0b7889, 'name': SearchDatastore_Task, 'duration_secs': 0.008609} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.367822] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9148114c-115f-45bf-8e60-2cfe999a67d6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.373192] env[62385]: DEBUG oslo_vmware.api [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 960.373192] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c02435-3c9f-a8cf-53c4-70533a01a56c" [ 960.373192] env[62385]: _type = "Task" [ 960.373192] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.381035] env[62385]: DEBUG oslo_vmware.api [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c02435-3c9f-a8cf-53c4-70533a01a56c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.619676] env[62385]: DEBUG oslo_vmware.api [None req-a35978b7-46d7-4e06-9dbf-0d023b5874ca tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206190, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.673134] env[62385]: DEBUG oslo_vmware.api [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Task: {'id': task-1206191, 'name': PowerOnVM_Task, 'duration_secs': 0.433405} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.673609] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] [instance: 104dea2c-6977-4683-bc32-05131858f562] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 960.673945] env[62385]: INFO nova.compute.manager [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] [instance: 104dea2c-6977-4683-bc32-05131858f562] Took 8.17 seconds to spawn the instance on the hypervisor. [ 960.674625] env[62385]: DEBUG nova.compute.manager [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] [instance: 104dea2c-6977-4683-bc32-05131858f562] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 960.674967] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cee42a4-a1ec-4690-8c10-59742562c538 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.781517] env[62385]: DEBUG nova.network.neutron [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Successfully updated port: a9513285-4023-4f3e-b90d-3eb507955f13 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 960.849937] env[62385]: DEBUG oslo_concurrency.lockutils [None req-782e1513-c3dc-4b59-8b0f-73e617258048 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.801s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.856093] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.849s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.862853] env[62385]: INFO nova.compute.claims [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 960.887950] env[62385]: DEBUG oslo_vmware.api [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c02435-3c9f-a8cf-53c4-70533a01a56c, 'name': SearchDatastore_Task, 'duration_secs': 0.009302} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.888267] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.888897] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] fd0b59c2-e38d-4ecc-b32f-72f07d555ffd/fd0b59c2-e38d-4ecc-b32f-72f07d555ffd.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 960.889053] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6883fc7b-dd74-49da-88ba-642f08781c33 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.893250] env[62385]: INFO nova.scheduler.client.report [None req-782e1513-c3dc-4b59-8b0f-73e617258048 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Deleted allocations for instance f1d33401-35f5-4d79-abb4-26dc6faa784e [ 960.895932] env[62385]: DEBUG oslo_vmware.api [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 960.895932] env[62385]: value = "task-1206192" [ 960.895932] env[62385]: _type = "Task" [ 960.895932] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.906557] env[62385]: DEBUG oslo_vmware.api [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206192, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.947684] env[62385]: DEBUG nova.compute.manager [req-6cf8e7cf-c1cf-4eec-8e8e-ed8a698cdbe8 req-45eb3409-83db-4540-93d1-65e817aff2ed service nova] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Received event network-vif-plugged-a9513285-4023-4f3e-b90d-3eb507955f13 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 960.947919] env[62385]: DEBUG oslo_concurrency.lockutils [req-6cf8e7cf-c1cf-4eec-8e8e-ed8a698cdbe8 req-45eb3409-83db-4540-93d1-65e817aff2ed service nova] Acquiring lock "df020d6b-3fab-4599-a342-47c7833b4240-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.948594] env[62385]: DEBUG oslo_concurrency.lockutils [req-6cf8e7cf-c1cf-4eec-8e8e-ed8a698cdbe8 req-45eb3409-83db-4540-93d1-65e817aff2ed service nova] Lock "df020d6b-3fab-4599-a342-47c7833b4240-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.948785] env[62385]: DEBUG oslo_concurrency.lockutils [req-6cf8e7cf-c1cf-4eec-8e8e-ed8a698cdbe8 req-45eb3409-83db-4540-93d1-65e817aff2ed service nova] Lock "df020d6b-3fab-4599-a342-47c7833b4240-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.948964] env[62385]: DEBUG nova.compute.manager [req-6cf8e7cf-c1cf-4eec-8e8e-ed8a698cdbe8 req-45eb3409-83db-4540-93d1-65e817aff2ed service nova] [instance: df020d6b-3fab-4599-a342-47c7833b4240] No waiting events found dispatching network-vif-plugged-a9513285-4023-4f3e-b90d-3eb507955f13 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 960.949793] env[62385]: WARNING nova.compute.manager [req-6cf8e7cf-c1cf-4eec-8e8e-ed8a698cdbe8 req-45eb3409-83db-4540-93d1-65e817aff2ed service nova] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Received unexpected event network-vif-plugged-a9513285-4023-4f3e-b90d-3eb507955f13 for instance with vm_state building and task_state spawning. [ 960.996544] env[62385]: DEBUG nova.network.neutron [req-12e381ce-9760-4b6c-acea-8bcedbf41da3 req-bedf00cd-ec98-44ed-854e-918e4f90d728 service nova] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Updated VIF entry in instance network info cache for port 091890dd-d345-4cee-8b38-f7f812092c5f. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 960.997041] env[62385]: DEBUG nova.network.neutron [req-12e381ce-9760-4b6c-acea-8bcedbf41da3 req-bedf00cd-ec98-44ed-854e-918e4f90d728 service nova] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Updating instance_info_cache with network_info: [{"id": "2b05b2b6-d4ce-4eb7-9d65-730d4dbad51c", "address": "fa:16:3e:ec:d7:16", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b05b2b6-d4", "ovs_interfaceid": "2b05b2b6-d4ce-4eb7-9d65-730d4dbad51c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c053ce8b-fc9f-4c48-9073-cdd753630700", "address": "fa:16:3e:9c:f6:87", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc053ce8b-fc", "ovs_interfaceid": "c053ce8b-fc9f-4c48-9073-cdd753630700", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "091890dd-d345-4cee-8b38-f7f812092c5f", "address": "fa:16:3e:b3:d9:f5", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap091890dd-d3", "ovs_interfaceid": "091890dd-d345-4cee-8b38-f7f812092c5f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.115054] env[62385]: DEBUG oslo_vmware.api [None req-a35978b7-46d7-4e06-9dbf-0d023b5874ca tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206190, 'name': ReconfigVM_Task, 'duration_secs': 0.62729} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.115534] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a35978b7-46d7-4e06-9dbf-0d023b5874ca tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Releasing lock "f937ef7b-ab39-4d8a-9577-7faabc652aa5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.115978] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a35978b7-46d7-4e06-9dbf-0d023b5874ca tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Reconfigured VM to attach interface {{(pid=62385) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 961.191231] env[62385]: DEBUG nova.network.neutron [req-f57c1736-6b38-4e21-8135-c2ef0cf0f13f req-b5349e81-495e-401f-9f34-c1f01d85cf6a service nova] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Updated VIF entry in instance network info cache for port 1b04305d-1213-4f55-9b69-6f15de71f576. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 961.191596] env[62385]: DEBUG nova.network.neutron [req-f57c1736-6b38-4e21-8135-c2ef0cf0f13f req-b5349e81-495e-401f-9f34-c1f01d85cf6a service nova] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Updating instance_info_cache with network_info: [{"id": "1b04305d-1213-4f55-9b69-6f15de71f576", "address": "fa:16:3e:85:9d:e7", "network": {"id": "afae6b07-d08c-413f-aa73-b545db1bb766", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1654965885-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "93111f5ba17c47de9691f90b99e03aa8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b04305d-12", "ovs_interfaceid": "1b04305d-1213-4f55-9b69-6f15de71f576", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.198130] env[62385]: INFO nova.compute.manager [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] [instance: 104dea2c-6977-4683-bc32-05131858f562] Took 24.28 seconds to build instance. [ 961.285550] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "refresh_cache-df020d6b-3fab-4599-a342-47c7833b4240" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.285664] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquired lock "refresh_cache-df020d6b-3fab-4599-a342-47c7833b4240" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.285869] env[62385]: DEBUG nova.network.neutron [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 961.407699] env[62385]: DEBUG oslo_vmware.api [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206192, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.410977] env[62385]: DEBUG oslo_concurrency.lockutils [None req-782e1513-c3dc-4b59-8b0f-73e617258048 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "f1d33401-35f5-4d79-abb4-26dc6faa784e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.240s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.499482] env[62385]: DEBUG oslo_concurrency.lockutils [req-12e381ce-9760-4b6c-acea-8bcedbf41da3 req-bedf00cd-ec98-44ed-854e-918e4f90d728 service nova] Releasing lock "refresh_cache-f937ef7b-ab39-4d8a-9577-7faabc652aa5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.620416] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a35978b7-46d7-4e06-9dbf-0d023b5874ca tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "interface-f937ef7b-ab39-4d8a-9577-7faabc652aa5-091890dd-d345-4cee-8b38-f7f812092c5f" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.202s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.695787] env[62385]: DEBUG oslo_concurrency.lockutils [req-f57c1736-6b38-4e21-8135-c2ef0cf0f13f req-b5349e81-495e-401f-9f34-c1f01d85cf6a service nova] Releasing lock "refresh_cache-5daf0a99-0c2a-40d8-afc7-1998e21e32d8" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.696647] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f3ef670e-47bf-4c58-82b8-dbcfba214330 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Lock "104dea2c-6977-4683-bc32-05131858f562" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.799s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.767113] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bebcac0a-2c03-43c8-925b-3f44500f4631 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Acquiring lock "104dea2c-6977-4683-bc32-05131858f562" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.767500] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bebcac0a-2c03-43c8-925b-3f44500f4631 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Lock "104dea2c-6977-4683-bc32-05131858f562" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.767848] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bebcac0a-2c03-43c8-925b-3f44500f4631 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Acquiring lock "104dea2c-6977-4683-bc32-05131858f562-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.768145] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bebcac0a-2c03-43c8-925b-3f44500f4631 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Lock "104dea2c-6977-4683-bc32-05131858f562-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.768409] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bebcac0a-2c03-43c8-925b-3f44500f4631 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Lock "104dea2c-6977-4683-bc32-05131858f562-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.771394] env[62385]: INFO nova.compute.manager [None req-bebcac0a-2c03-43c8-925b-3f44500f4631 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] [instance: 104dea2c-6977-4683-bc32-05131858f562] Terminating instance [ 961.773975] env[62385]: DEBUG nova.compute.manager [None req-bebcac0a-2c03-43c8-925b-3f44500f4631 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] [instance: 104dea2c-6977-4683-bc32-05131858f562] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 961.774280] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bebcac0a-2c03-43c8-925b-3f44500f4631 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] [instance: 104dea2c-6977-4683-bc32-05131858f562] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 961.775470] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-153e630d-d0eb-4fce-9812-69bffef93e4b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.785833] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bebcac0a-2c03-43c8-925b-3f44500f4631 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] [instance: 104dea2c-6977-4683-bc32-05131858f562] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 961.786273] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-abfcb117-ad01-4194-a8b5-791404132816 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.795022] env[62385]: DEBUG oslo_vmware.api [None req-bebcac0a-2c03-43c8-925b-3f44500f4631 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Waiting for the task: (returnval){ [ 961.795022] env[62385]: value = "task-1206193" [ 961.795022] env[62385]: _type = "Task" [ 961.795022] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.804352] env[62385]: DEBUG oslo_vmware.api [None req-bebcac0a-2c03-43c8-925b-3f44500f4631 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Task: {'id': task-1206193, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.822602] env[62385]: DEBUG nova.network.neutron [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 961.841963] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f6add1d4-a93d-4ad5-bbb2-9f5005aeead7 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquiring lock "33387505-c576-488b-8c9c-b064fe81a7d7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.842326] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f6add1d4-a93d-4ad5-bbb2-9f5005aeead7 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "33387505-c576-488b-8c9c-b064fe81a7d7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.842622] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f6add1d4-a93d-4ad5-bbb2-9f5005aeead7 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquiring lock "33387505-c576-488b-8c9c-b064fe81a7d7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.842903] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f6add1d4-a93d-4ad5-bbb2-9f5005aeead7 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "33387505-c576-488b-8c9c-b064fe81a7d7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.843170] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f6add1d4-a93d-4ad5-bbb2-9f5005aeead7 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "33387505-c576-488b-8c9c-b064fe81a7d7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.846043] env[62385]: INFO nova.compute.manager [None req-f6add1d4-a93d-4ad5-bbb2-9f5005aeead7 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Terminating instance [ 961.848378] env[62385]: DEBUG nova.compute.manager [None req-f6add1d4-a93d-4ad5-bbb2-9f5005aeead7 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 961.848672] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-f6add1d4-a93d-4ad5-bbb2-9f5005aeead7 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 961.849867] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53338abd-d2de-4026-a7a9-e13810a3464d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.859731] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6add1d4-a93d-4ad5-bbb2-9f5005aeead7 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 961.859953] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-79aeb2d2-28f9-48d1-b56b-eff00aff9d9c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.866860] env[62385]: DEBUG oslo_vmware.api [None req-f6add1d4-a93d-4ad5-bbb2-9f5005aeead7 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 961.866860] env[62385]: value = "task-1206194" [ 961.866860] env[62385]: _type = "Task" [ 961.866860] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.881579] env[62385]: DEBUG oslo_vmware.api [None req-f6add1d4-a93d-4ad5-bbb2-9f5005aeead7 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206194, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.905826] env[62385]: DEBUG oslo_vmware.api [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206192, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.961180] env[62385]: DEBUG nova.network.neutron [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Updating instance_info_cache with network_info: [{"id": "a9513285-4023-4f3e-b90d-3eb507955f13", "address": "fa:16:3e:30:02:c0", "network": {"id": "30c4e78b-6a64-47dc-aaaa-a210160c9f23", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1809587656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "105150c170d74e268e82ab894e9bf0d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10ff2092-e8eb-4768-ad4a-65a80560b447", "external-id": "nsx-vlan-transportzone-845", "segmentation_id": 845, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa9513285-40", "ovs_interfaceid": "a9513285-4023-4f3e-b90d-3eb507955f13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.064927] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c57b2277-d07a-47df-9ac3-bd8b08043b93 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.073265] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-749954b0-bfe8-4871-b84e-08dbac7b423f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.103158] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ff3d91c-df51-423a-95ee-22e693b5a86a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.110916] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f77ab908-47b5-480f-942b-848402892083 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.125738] env[62385]: DEBUG nova.compute.provider_tree [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 962.305018] env[62385]: DEBUG oslo_vmware.api [None req-bebcac0a-2c03-43c8-925b-3f44500f4631 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Task: {'id': task-1206193, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.367879] env[62385]: DEBUG nova.compute.manager [req-5a499a57-5858-4bb8-a816-0bfe30e5ca1b req-838d6ca9-d356-488a-a656-f4a3fb3fa1b8 service nova] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Received event network-changed-a9513285-4023-4f3e-b90d-3eb507955f13 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 962.368045] env[62385]: DEBUG nova.compute.manager [req-5a499a57-5858-4bb8-a816-0bfe30e5ca1b req-838d6ca9-d356-488a-a656-f4a3fb3fa1b8 service nova] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Refreshing instance network info cache due to event network-changed-a9513285-4023-4f3e-b90d-3eb507955f13. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 962.368258] env[62385]: DEBUG oslo_concurrency.lockutils [req-5a499a57-5858-4bb8-a816-0bfe30e5ca1b req-838d6ca9-d356-488a-a656-f4a3fb3fa1b8 service nova] Acquiring lock "refresh_cache-df020d6b-3fab-4599-a342-47c7833b4240" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.378210] env[62385]: DEBUG oslo_vmware.api [None req-f6add1d4-a93d-4ad5-bbb2-9f5005aeead7 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206194, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.408525] env[62385]: DEBUG oslo_vmware.api [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206192, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.467189] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Releasing lock "refresh_cache-df020d6b-3fab-4599-a342-47c7833b4240" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.467649] env[62385]: DEBUG nova.compute.manager [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Instance network_info: |[{"id": "a9513285-4023-4f3e-b90d-3eb507955f13", "address": "fa:16:3e:30:02:c0", "network": {"id": "30c4e78b-6a64-47dc-aaaa-a210160c9f23", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1809587656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "105150c170d74e268e82ab894e9bf0d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10ff2092-e8eb-4768-ad4a-65a80560b447", "external-id": "nsx-vlan-transportzone-845", "segmentation_id": 845, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa9513285-40", "ovs_interfaceid": "a9513285-4023-4f3e-b90d-3eb507955f13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 962.468007] env[62385]: DEBUG oslo_concurrency.lockutils [req-5a499a57-5858-4bb8-a816-0bfe30e5ca1b req-838d6ca9-d356-488a-a656-f4a3fb3fa1b8 service nova] Acquired lock "refresh_cache-df020d6b-3fab-4599-a342-47c7833b4240" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.468216] env[62385]: DEBUG nova.network.neutron [req-5a499a57-5858-4bb8-a816-0bfe30e5ca1b req-838d6ca9-d356-488a-a656-f4a3fb3fa1b8 service nova] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Refreshing network info cache for port a9513285-4023-4f3e-b90d-3eb507955f13 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 962.469505] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:30:02:c0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '10ff2092-e8eb-4768-ad4a-65a80560b447', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a9513285-4023-4f3e-b90d-3eb507955f13', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 962.477232] env[62385]: DEBUG oslo.service.loopingcall [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 962.477345] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 962.477545] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3714ac3c-71c1-4dc0-b3d4-3093ab239d88 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.498059] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 962.498059] env[62385]: value = "task-1206195" [ 962.498059] env[62385]: _type = "Task" [ 962.498059] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.506219] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206195, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.627839] env[62385]: DEBUG nova.scheduler.client.report [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 962.806277] env[62385]: DEBUG oslo_vmware.api [None req-bebcac0a-2c03-43c8-925b-3f44500f4631 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Task: {'id': task-1206193, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.880624] env[62385]: DEBUG oslo_vmware.api [None req-f6add1d4-a93d-4ad5-bbb2-9f5005aeead7 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206194, 'name': PowerOffVM_Task, 'duration_secs': 0.976587} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.880932] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6add1d4-a93d-4ad5-bbb2-9f5005aeead7 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 962.881204] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-f6add1d4-a93d-4ad5-bbb2-9f5005aeead7 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 962.881548] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bf80630b-2f6b-40a3-987e-bae9e6a133f3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.907759] env[62385]: DEBUG oslo_vmware.api [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206192, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.687691} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.908043] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] fd0b59c2-e38d-4ecc-b32f-72f07d555ffd/fd0b59c2-e38d-4ecc-b32f-72f07d555ffd.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 962.908380] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 962.908738] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6854a8e9-9470-4801-9b0a-b1f8c2e6bd90 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.915134] env[62385]: DEBUG oslo_vmware.api [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 962.915134] env[62385]: value = "task-1206197" [ 962.915134] env[62385]: _type = "Task" [ 962.915134] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.922888] env[62385]: DEBUG oslo_vmware.api [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206197, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.957699] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-f6add1d4-a93d-4ad5-bbb2-9f5005aeead7 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 962.957991] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-f6add1d4-a93d-4ad5-bbb2-9f5005aeead7 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Deleting contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 962.958185] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6add1d4-a93d-4ad5-bbb2-9f5005aeead7 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Deleting the datastore file [datastore2] 33387505-c576-488b-8c9c-b064fe81a7d7 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 962.958477] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-80778316-dbde-4e1f-8151-f8e3e9723b3d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.964801] env[62385]: DEBUG oslo_vmware.api [None req-f6add1d4-a93d-4ad5-bbb2-9f5005aeead7 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for the task: (returnval){ [ 962.964801] env[62385]: value = "task-1206198" [ 962.964801] env[62385]: _type = "Task" [ 962.964801] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.973539] env[62385]: DEBUG oslo_vmware.api [None req-f6add1d4-a93d-4ad5-bbb2-9f5005aeead7 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206198, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.007959] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206195, 'name': CreateVM_Task, 'duration_secs': 0.506628} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.010153] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 963.010859] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.011059] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.011381] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 963.011917] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e7aea881-8837-492e-a30c-bb25d611a7da {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.016734] env[62385]: DEBUG oslo_vmware.api [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 963.016734] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]525f51d0-90c0-d623-02e8-221639724176" [ 963.016734] env[62385]: _type = "Task" [ 963.016734] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.024061] env[62385]: DEBUG oslo_vmware.api [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]525f51d0-90c0-d623-02e8-221639724176, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.134755] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.279s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.135297] env[62385]: DEBUG nova.compute.manager [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 963.138288] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c319277e-63ff-46fd-a975-49b0644f34a8 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.614s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.138446] env[62385]: DEBUG nova.objects.instance [None req-c319277e-63ff-46fd-a975-49b0644f34a8 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lazy-loading 'resources' on Instance uuid df559cb8-555c-496b-bce3-0981be3ef65f {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 963.224422] env[62385]: DEBUG nova.network.neutron [req-5a499a57-5858-4bb8-a816-0bfe30e5ca1b req-838d6ca9-d356-488a-a656-f4a3fb3fa1b8 service nova] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Updated VIF entry in instance network info cache for port a9513285-4023-4f3e-b90d-3eb507955f13. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 963.224780] env[62385]: DEBUG nova.network.neutron [req-5a499a57-5858-4bb8-a816-0bfe30e5ca1b req-838d6ca9-d356-488a-a656-f4a3fb3fa1b8 service nova] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Updating instance_info_cache with network_info: [{"id": "a9513285-4023-4f3e-b90d-3eb507955f13", "address": "fa:16:3e:30:02:c0", "network": {"id": "30c4e78b-6a64-47dc-aaaa-a210160c9f23", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1809587656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "105150c170d74e268e82ab894e9bf0d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10ff2092-e8eb-4768-ad4a-65a80560b447", "external-id": "nsx-vlan-transportzone-845", "segmentation_id": 845, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa9513285-40", "ovs_interfaceid": "a9513285-4023-4f3e-b90d-3eb507955f13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.273816] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a0e27224-d55b-467b-b751-b4feec4d5360 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "interface-f937ef7b-ab39-4d8a-9577-7faabc652aa5-c053ce8b-fc9f-4c48-9073-cdd753630700" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.274216] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a0e27224-d55b-467b-b751-b4feec4d5360 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "interface-f937ef7b-ab39-4d8a-9577-7faabc652aa5-c053ce8b-fc9f-4c48-9073-cdd753630700" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.305453] env[62385]: DEBUG oslo_vmware.api [None req-bebcac0a-2c03-43c8-925b-3f44500f4631 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Task: {'id': task-1206193, 'name': PowerOffVM_Task, 'duration_secs': 1.051757} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.305719] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bebcac0a-2c03-43c8-925b-3f44500f4631 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] [instance: 104dea2c-6977-4683-bc32-05131858f562] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 963.305898] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bebcac0a-2c03-43c8-925b-3f44500f4631 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] [instance: 104dea2c-6977-4683-bc32-05131858f562] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 963.306175] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6b9faa30-4a64-42fd-996e-599a183c7d05 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.373255] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bebcac0a-2c03-43c8-925b-3f44500f4631 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] [instance: 104dea2c-6977-4683-bc32-05131858f562] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 963.373496] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bebcac0a-2c03-43c8-925b-3f44500f4631 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] [instance: 104dea2c-6977-4683-bc32-05131858f562] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 963.373686] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bebcac0a-2c03-43c8-925b-3f44500f4631 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Deleting the datastore file [datastore1] 104dea2c-6977-4683-bc32-05131858f562 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 963.373963] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e8ac2e74-69ba-4dcb-ad6b-b124d9005cd1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.380766] env[62385]: DEBUG oslo_vmware.api [None req-bebcac0a-2c03-43c8-925b-3f44500f4631 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Waiting for the task: (returnval){ [ 963.380766] env[62385]: value = "task-1206200" [ 963.380766] env[62385]: _type = "Task" [ 963.380766] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.389270] env[62385]: DEBUG oslo_vmware.api [None req-bebcac0a-2c03-43c8-925b-3f44500f4631 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Task: {'id': task-1206200, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.423370] env[62385]: DEBUG oslo_vmware.api [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206197, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.05537} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.423599] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 963.424358] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-532ef137-e8bc-48e1-acd5-7ec30f004b07 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.445929] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Reconfiguring VM instance instance-0000005f to attach disk [datastore1] fd0b59c2-e38d-4ecc-b32f-72f07d555ffd/fd0b59c2-e38d-4ecc-b32f-72f07d555ffd.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 963.446197] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ace0b74e-04bf-4424-b50a-9e1604405cb5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.464340] env[62385]: DEBUG oslo_vmware.api [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 963.464340] env[62385]: value = "task-1206201" [ 963.464340] env[62385]: _type = "Task" [ 963.464340] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.474708] env[62385]: DEBUG oslo_vmware.api [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206201, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.477649] env[62385]: DEBUG oslo_vmware.api [None req-f6add1d4-a93d-4ad5-bbb2-9f5005aeead7 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Task: {'id': task-1206198, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.325731} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.477879] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6add1d4-a93d-4ad5-bbb2-9f5005aeead7 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 963.478103] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-f6add1d4-a93d-4ad5-bbb2-9f5005aeead7 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Deleted contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 963.478379] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-f6add1d4-a93d-4ad5-bbb2-9f5005aeead7 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 963.478579] env[62385]: INFO nova.compute.manager [None req-f6add1d4-a93d-4ad5-bbb2-9f5005aeead7 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Took 1.63 seconds to destroy the instance on the hypervisor. [ 963.478819] env[62385]: DEBUG oslo.service.loopingcall [None req-f6add1d4-a93d-4ad5-bbb2-9f5005aeead7 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 963.479015] env[62385]: DEBUG nova.compute.manager [-] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 963.479122] env[62385]: DEBUG nova.network.neutron [-] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 963.529225] env[62385]: DEBUG oslo_vmware.api [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]525f51d0-90c0-d623-02e8-221639724176, 'name': SearchDatastore_Task, 'duration_secs': 0.047435} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.529584] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.529828] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 963.530078] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.530234] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.530420] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 963.530688] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d64679c7-4b1b-45a5-bee7-9ee04ab8aadb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.537788] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 963.537995] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 963.538689] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a79cc981-7dae-4f14-8b03-fc4d9b3309e6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.543838] env[62385]: DEBUG oslo_vmware.api [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 963.543838] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]526ba428-3c65-229d-0e7f-77e2476181a7" [ 963.543838] env[62385]: _type = "Task" [ 963.543838] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.551126] env[62385]: DEBUG oslo_vmware.api [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]526ba428-3c65-229d-0e7f-77e2476181a7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.641698] env[62385]: DEBUG nova.compute.utils [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 963.646095] env[62385]: DEBUG nova.compute.manager [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 963.646283] env[62385]: DEBUG nova.network.neutron [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 963.709483] env[62385]: DEBUG nova.policy [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e1200cdf75b04233b4c3f4626d4e27cf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aea48968d4464ac0ab01c998c059107d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 963.727678] env[62385]: DEBUG oslo_concurrency.lockutils [req-5a499a57-5858-4bb8-a816-0bfe30e5ca1b req-838d6ca9-d356-488a-a656-f4a3fb3fa1b8 service nova] Releasing lock "refresh_cache-df020d6b-3fab-4599-a342-47c7833b4240" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.778346] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a0e27224-d55b-467b-b751-b4feec4d5360 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "f937ef7b-ab39-4d8a-9577-7faabc652aa5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.778500] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a0e27224-d55b-467b-b751-b4feec4d5360 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquired lock "f937ef7b-ab39-4d8a-9577-7faabc652aa5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.779737] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d93239f0-98d7-4045-9177-5fe33e5c6769 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.806381] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5597da8-2f70-4e9d-9abb-c4c933cb9eca {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.834898] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a0e27224-d55b-467b-b751-b4feec4d5360 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Reconfiguring VM to detach interface {{(pid=62385) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 963.838020] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f6d4c02a-5327-4c7e-9e0b-35fa2c6c8b76 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.857472] env[62385]: DEBUG oslo_vmware.api [None req-a0e27224-d55b-467b-b751-b4feec4d5360 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 963.857472] env[62385]: value = "task-1206202" [ 963.857472] env[62385]: _type = "Task" [ 963.857472] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.865958] env[62385]: DEBUG oslo_vmware.api [None req-a0e27224-d55b-467b-b751-b4feec4d5360 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206202, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.891676] env[62385]: DEBUG oslo_vmware.api [None req-bebcac0a-2c03-43c8-925b-3f44500f4631 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Task: {'id': task-1206200, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146087} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.894183] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-bebcac0a-2c03-43c8-925b-3f44500f4631 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 963.894396] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bebcac0a-2c03-43c8-925b-3f44500f4631 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] [instance: 104dea2c-6977-4683-bc32-05131858f562] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 963.894557] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bebcac0a-2c03-43c8-925b-3f44500f4631 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] [instance: 104dea2c-6977-4683-bc32-05131858f562] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 963.894843] env[62385]: INFO nova.compute.manager [None req-bebcac0a-2c03-43c8-925b-3f44500f4631 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] [instance: 104dea2c-6977-4683-bc32-05131858f562] Took 2.12 seconds to destroy the instance on the hypervisor. [ 963.895123] env[62385]: DEBUG oslo.service.loopingcall [None req-bebcac0a-2c03-43c8-925b-3f44500f4631 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 963.895501] env[62385]: DEBUG nova.compute.manager [-] [instance: 104dea2c-6977-4683-bc32-05131858f562] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 963.895612] env[62385]: DEBUG nova.network.neutron [-] [instance: 104dea2c-6977-4683-bc32-05131858f562] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 963.919178] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64929021-0e83-4a43-9fc1-fb141d65aae9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.927064] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b797e3d-b211-421d-bf3f-1a92fc85a948 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.957079] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fa19059-7920-4915-93a7-41b9599039e7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.964245] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77e6b3da-7fd1-4f72-b72d-2d7a8c600e67 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.986027] env[62385]: DEBUG nova.compute.provider_tree [None req-c319277e-63ff-46fd-a975-49b0644f34a8 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 963.987235] env[62385]: DEBUG oslo_vmware.api [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206201, 'name': ReconfigVM_Task, 'duration_secs': 0.365355} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.987487] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Reconfigured VM instance instance-0000005f to attach disk [datastore1] fd0b59c2-e38d-4ecc-b32f-72f07d555ffd/fd0b59c2-e38d-4ecc-b32f-72f07d555ffd.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 963.988168] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7996f29b-1346-48ac-880e-a72123ca7968 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.996762] env[62385]: DEBUG oslo_vmware.api [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 963.996762] env[62385]: value = "task-1206203" [ 963.996762] env[62385]: _type = "Task" [ 963.996762] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.004883] env[62385]: DEBUG oslo_vmware.api [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206203, 'name': Rename_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.057150] env[62385]: DEBUG oslo_vmware.api [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]526ba428-3c65-229d-0e7f-77e2476181a7, 'name': SearchDatastore_Task, 'duration_secs': 0.007965} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.057990] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8f59e132-8674-4e6a-bc73-36db5c1b72b3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.063820] env[62385]: DEBUG oslo_vmware.api [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 964.063820] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]529825f7-a065-4225-2d15-53e47b761c08" [ 964.063820] env[62385]: _type = "Task" [ 964.063820] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.072386] env[62385]: DEBUG oslo_vmware.api [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]529825f7-a065-4225-2d15-53e47b761c08, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.147142] env[62385]: DEBUG nova.compute.manager [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 964.253891] env[62385]: DEBUG nova.network.neutron [-] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.351116] env[62385]: DEBUG nova.network.neutron [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Successfully created port: eef524f4-fc00-47f1-a485-dcf568c6d0e4 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 964.368295] env[62385]: DEBUG oslo_vmware.api [None req-a0e27224-d55b-467b-b751-b4feec4d5360 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206202, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.400404] env[62385]: DEBUG nova.compute.manager [req-82f71893-62ee-4685-a674-1212bd41dbfd req-ce642ae9-57cb-41a1-bde0-e02e5bbf4fdc service nova] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Received event network-vif-deleted-72d846e7-a597-4eb7-87ea-0d4309ceb42b {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 964.491485] env[62385]: DEBUG nova.scheduler.client.report [None req-c319277e-63ff-46fd-a975-49b0644f34a8 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 964.506437] env[62385]: DEBUG oslo_vmware.api [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206203, 'name': Rename_Task, 'duration_secs': 0.136268} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.506726] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 964.507033] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-19e22346-4735-47db-bb79-0e26c0aa64fc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.514711] env[62385]: DEBUG oslo_vmware.api [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 964.514711] env[62385]: value = "task-1206204" [ 964.514711] env[62385]: _type = "Task" [ 964.514711] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.522593] env[62385]: DEBUG oslo_vmware.api [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206204, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.575606] env[62385]: DEBUG oslo_vmware.api [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]529825f7-a065-4225-2d15-53e47b761c08, 'name': SearchDatastore_Task, 'duration_secs': 0.009956} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.575942] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.576612] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] df020d6b-3fab-4599-a342-47c7833b4240/df020d6b-3fab-4599-a342-47c7833b4240.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 964.577387] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-90f3e05b-5a46-49e2-9d7d-213821e279d9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.588098] env[62385]: DEBUG oslo_vmware.api [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 964.588098] env[62385]: value = "task-1206205" [ 964.588098] env[62385]: _type = "Task" [ 964.588098] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.599043] env[62385]: DEBUG oslo_vmware.api [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206205, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.755508] env[62385]: INFO nova.compute.manager [-] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Took 1.28 seconds to deallocate network for instance. [ 964.870760] env[62385]: DEBUG oslo_vmware.api [None req-a0e27224-d55b-467b-b751-b4feec4d5360 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206202, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.912196] env[62385]: DEBUG nova.network.neutron [-] [instance: 104dea2c-6977-4683-bc32-05131858f562] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.996831] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c319277e-63ff-46fd-a975-49b0644f34a8 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.858s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.000076] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.685s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.002499] env[62385]: INFO nova.compute.claims [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 965.022021] env[62385]: INFO nova.scheduler.client.report [None req-c319277e-63ff-46fd-a975-49b0644f34a8 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Deleted allocations for instance df559cb8-555c-496b-bce3-0981be3ef65f [ 965.028970] env[62385]: DEBUG oslo_vmware.api [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206204, 'name': PowerOnVM_Task, 'duration_secs': 0.459624} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.029264] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 965.029469] env[62385]: INFO nova.compute.manager [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Took 10.18 seconds to spawn the instance on the hypervisor. [ 965.029678] env[62385]: DEBUG nova.compute.manager [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 965.030480] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d4ba045-d85f-46d1-8f64-712f59d7f50f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.098861] env[62385]: DEBUG oslo_vmware.api [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206205, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.434258} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.098861] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] df020d6b-3fab-4599-a342-47c7833b4240/df020d6b-3fab-4599-a342-47c7833b4240.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 965.098861] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 965.099198] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7bd6512c-401f-4756-8567-dbb027e9dd88 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.105841] env[62385]: DEBUG oslo_vmware.api [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 965.105841] env[62385]: value = "task-1206206" [ 965.105841] env[62385]: _type = "Task" [ 965.105841] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.116304] env[62385]: DEBUG oslo_vmware.api [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206206, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.156317] env[62385]: DEBUG nova.compute.manager [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 965.183696] env[62385]: DEBUG nova.virt.hardware [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 965.184029] env[62385]: DEBUG nova.virt.hardware [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 965.184235] env[62385]: DEBUG nova.virt.hardware [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 965.184448] env[62385]: DEBUG nova.virt.hardware [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 965.184605] env[62385]: DEBUG nova.virt.hardware [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 965.184762] env[62385]: DEBUG nova.virt.hardware [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 965.184987] env[62385]: DEBUG nova.virt.hardware [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 965.185644] env[62385]: DEBUG nova.virt.hardware [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 965.185644] env[62385]: DEBUG nova.virt.hardware [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 965.185644] env[62385]: DEBUG nova.virt.hardware [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 965.185843] env[62385]: DEBUG nova.virt.hardware [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 965.186586] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efb7eb0d-7e8f-4d5d-b734-32c827d59d38 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.195237] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0a19dbb-7382-4018-96c0-df84a59236c5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.266851] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f6add1d4-a93d-4ad5-bbb2-9f5005aeead7 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.369505] env[62385]: DEBUG oslo_vmware.api [None req-a0e27224-d55b-467b-b751-b4feec4d5360 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206202, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.416131] env[62385]: INFO nova.compute.manager [-] [instance: 104dea2c-6977-4683-bc32-05131858f562] Took 1.52 seconds to deallocate network for instance. [ 965.528965] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c319277e-63ff-46fd-a975-49b0644f34a8 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "df559cb8-555c-496b-bce3-0981be3ef65f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.060s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.548924] env[62385]: INFO nova.compute.manager [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Took 23.82 seconds to build instance. [ 965.620179] env[62385]: DEBUG oslo_vmware.api [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206206, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062287} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.621035] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 965.621657] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f7a1250-5825-4adc-9c75-192fac7a8b26 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.655999] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] df020d6b-3fab-4599-a342-47c7833b4240/df020d6b-3fab-4599-a342-47c7833b4240.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 965.656905] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-827a09b3-cb16-4077-953c-7de2c199468d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.678039] env[62385]: DEBUG oslo_vmware.api [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 965.678039] env[62385]: value = "task-1206207" [ 965.678039] env[62385]: _type = "Task" [ 965.678039] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.686151] env[62385]: DEBUG oslo_vmware.api [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206207, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.873874] env[62385]: DEBUG oslo_vmware.api [None req-a0e27224-d55b-467b-b751-b4feec4d5360 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206202, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.922758] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bebcac0a-2c03-43c8-925b-3f44500f4631 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.050928] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9aa8c069-9911-418a-958d-497b91aa514c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "fd0b59c2-e38d-4ecc-b32f-72f07d555ffd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.333s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.112348] env[62385]: DEBUG nova.network.neutron [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Successfully updated port: eef524f4-fc00-47f1-a485-dcf568c6d0e4 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 966.190472] env[62385]: DEBUG oslo_vmware.api [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206207, 'name': ReconfigVM_Task, 'duration_secs': 0.279642} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.190837] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Reconfigured VM instance instance-00000060 to attach disk [datastore2] df020d6b-3fab-4599-a342-47c7833b4240/df020d6b-3fab-4599-a342-47c7833b4240.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 966.191499] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3d5f50da-67b0-4eee-9453-4b5d9adce5bb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.198526] env[62385]: DEBUG oslo_vmware.api [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 966.198526] env[62385]: value = "task-1206208" [ 966.198526] env[62385]: _type = "Task" [ 966.198526] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.208018] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02b71b30-86da-4b1d-a201-64feff860bf0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.213138] env[62385]: DEBUG oslo_vmware.api [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206208, 'name': Rename_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.217627] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-196e4d80-ba9b-49ae-8a5f-43b4bee9e015 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.249082] env[62385]: INFO nova.compute.manager [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Rebuilding instance [ 966.251947] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c37abbb-413b-41be-9f7e-46f265650bb6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.259605] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c14c0d66-a26c-47ff-a1e6-2dfd499d2546 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.275661] env[62385]: DEBUG nova.compute.provider_tree [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 966.280459] env[62385]: DEBUG oslo_concurrency.lockutils [None req-89a795a6-938d-4015-8160-ceaccf5f0691 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "9b2f8292-9d89-407e-96c5-195ee398cc7b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.280705] env[62385]: DEBUG oslo_concurrency.lockutils [None req-89a795a6-938d-4015-8160-ceaccf5f0691 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "9b2f8292-9d89-407e-96c5-195ee398cc7b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.280925] env[62385]: DEBUG oslo_concurrency.lockutils [None req-89a795a6-938d-4015-8160-ceaccf5f0691 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "9b2f8292-9d89-407e-96c5-195ee398cc7b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.281223] env[62385]: DEBUG oslo_concurrency.lockutils [None req-89a795a6-938d-4015-8160-ceaccf5f0691 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "9b2f8292-9d89-407e-96c5-195ee398cc7b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.281422] env[62385]: DEBUG oslo_concurrency.lockutils [None req-89a795a6-938d-4015-8160-ceaccf5f0691 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "9b2f8292-9d89-407e-96c5-195ee398cc7b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.283553] env[62385]: INFO nova.compute.manager [None req-89a795a6-938d-4015-8160-ceaccf5f0691 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Terminating instance [ 966.285688] env[62385]: DEBUG nova.compute.manager [None req-89a795a6-938d-4015-8160-ceaccf5f0691 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 966.285868] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-89a795a6-938d-4015-8160-ceaccf5f0691 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 966.286694] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73fa6f57-8c0e-420b-9f15-1fd960ba640d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.294970] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-89a795a6-938d-4015-8160-ceaccf5f0691 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 966.297238] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0446a9f2-ac23-4553-a012-80c4cf6d26fc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.300959] env[62385]: DEBUG nova.compute.manager [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 966.302102] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3edb26ff-8cb9-46cd-86d1-31402ce636ef {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.306328] env[62385]: DEBUG oslo_vmware.api [None req-89a795a6-938d-4015-8160-ceaccf5f0691 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 966.306328] env[62385]: value = "task-1206209" [ 966.306328] env[62385]: _type = "Task" [ 966.306328] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.318583] env[62385]: DEBUG oslo_vmware.api [None req-89a795a6-938d-4015-8160-ceaccf5f0691 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206209, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.371754] env[62385]: DEBUG oslo_vmware.api [None req-a0e27224-d55b-467b-b751-b4feec4d5360 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206202, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.473149] env[62385]: DEBUG nova.compute.manager [req-f1419ec6-f4b5-4a84-9ac8-604846a54cfb req-a85029a7-f506-4a96-9e88-cb526c8f10d7 service nova] [instance: 104dea2c-6977-4683-bc32-05131858f562] Received event network-vif-deleted-dc5f8505-529f-458b-9f90-58f310ea7ad0 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 966.473149] env[62385]: DEBUG nova.compute.manager [req-f1419ec6-f4b5-4a84-9ac8-604846a54cfb req-a85029a7-f506-4a96-9e88-cb526c8f10d7 service nova] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Received event network-vif-plugged-eef524f4-fc00-47f1-a485-dcf568c6d0e4 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 966.473635] env[62385]: DEBUG oslo_concurrency.lockutils [req-f1419ec6-f4b5-4a84-9ac8-604846a54cfb req-a85029a7-f506-4a96-9e88-cb526c8f10d7 service nova] Acquiring lock "b1e0baf3-643c-49c2-8a80-8de07d8527e7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.473903] env[62385]: DEBUG oslo_concurrency.lockutils [req-f1419ec6-f4b5-4a84-9ac8-604846a54cfb req-a85029a7-f506-4a96-9e88-cb526c8f10d7 service nova] Lock "b1e0baf3-643c-49c2-8a80-8de07d8527e7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.474174] env[62385]: DEBUG oslo_concurrency.lockutils [req-f1419ec6-f4b5-4a84-9ac8-604846a54cfb req-a85029a7-f506-4a96-9e88-cb526c8f10d7 service nova] Lock "b1e0baf3-643c-49c2-8a80-8de07d8527e7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.474383] env[62385]: DEBUG nova.compute.manager [req-f1419ec6-f4b5-4a84-9ac8-604846a54cfb req-a85029a7-f506-4a96-9e88-cb526c8f10d7 service nova] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] No waiting events found dispatching network-vif-plugged-eef524f4-fc00-47f1-a485-dcf568c6d0e4 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 966.474595] env[62385]: WARNING nova.compute.manager [req-f1419ec6-f4b5-4a84-9ac8-604846a54cfb req-a85029a7-f506-4a96-9e88-cb526c8f10d7 service nova] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Received unexpected event network-vif-plugged-eef524f4-fc00-47f1-a485-dcf568c6d0e4 for instance with vm_state building and task_state spawning. [ 966.474799] env[62385]: DEBUG nova.compute.manager [req-f1419ec6-f4b5-4a84-9ac8-604846a54cfb req-a85029a7-f506-4a96-9e88-cb526c8f10d7 service nova] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Received event network-changed-eef524f4-fc00-47f1-a485-dcf568c6d0e4 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 966.474972] env[62385]: DEBUG nova.compute.manager [req-f1419ec6-f4b5-4a84-9ac8-604846a54cfb req-a85029a7-f506-4a96-9e88-cb526c8f10d7 service nova] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Refreshing instance network info cache due to event network-changed-eef524f4-fc00-47f1-a485-dcf568c6d0e4. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 966.475206] env[62385]: DEBUG oslo_concurrency.lockutils [req-f1419ec6-f4b5-4a84-9ac8-604846a54cfb req-a85029a7-f506-4a96-9e88-cb526c8f10d7 service nova] Acquiring lock "refresh_cache-b1e0baf3-643c-49c2-8a80-8de07d8527e7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.475375] env[62385]: DEBUG oslo_concurrency.lockutils [req-f1419ec6-f4b5-4a84-9ac8-604846a54cfb req-a85029a7-f506-4a96-9e88-cb526c8f10d7 service nova] Acquired lock "refresh_cache-b1e0baf3-643c-49c2-8a80-8de07d8527e7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.475572] env[62385]: DEBUG nova.network.neutron [req-f1419ec6-f4b5-4a84-9ac8-604846a54cfb req-a85029a7-f506-4a96-9e88-cb526c8f10d7 service nova] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Refreshing network info cache for port eef524f4-fc00-47f1-a485-dcf568c6d0e4 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 966.621707] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "refresh_cache-b1e0baf3-643c-49c2-8a80-8de07d8527e7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.708580] env[62385]: DEBUG oslo_vmware.api [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206208, 'name': Rename_Task, 'duration_secs': 0.179946} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.708862] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 966.709121] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e3a6ca1c-f096-4a82-afd3-ec088d260314 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.715412] env[62385]: DEBUG oslo_vmware.api [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 966.715412] env[62385]: value = "task-1206210" [ 966.715412] env[62385]: _type = "Task" [ 966.715412] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.722931] env[62385]: DEBUG oslo_vmware.api [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206210, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.784978] env[62385]: DEBUG nova.scheduler.client.report [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 966.815077] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 966.818202] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5a7c5ca1-5ad2-4c61-b12a-0e139cfea88e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.819745] env[62385]: DEBUG oslo_vmware.api [None req-89a795a6-938d-4015-8160-ceaccf5f0691 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206209, 'name': PowerOffVM_Task, 'duration_secs': 0.212798} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.820131] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-89a795a6-938d-4015-8160-ceaccf5f0691 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 966.820208] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-89a795a6-938d-4015-8160-ceaccf5f0691 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 966.820935] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3a2b768f-5430-4573-b8ec-b70e8dc33ec6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.824502] env[62385]: DEBUG oslo_vmware.api [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 966.824502] env[62385]: value = "task-1206211" [ 966.824502] env[62385]: _type = "Task" [ 966.824502] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.834257] env[62385]: DEBUG oslo_vmware.api [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206211, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.869931] env[62385]: DEBUG oslo_vmware.api [None req-a0e27224-d55b-467b-b751-b4feec4d5360 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206202, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.885192] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-89a795a6-938d-4015-8160-ceaccf5f0691 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 966.885662] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-89a795a6-938d-4015-8160-ceaccf5f0691 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 966.885662] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-89a795a6-938d-4015-8160-ceaccf5f0691 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Deleting the datastore file [datastore1] 9b2f8292-9d89-407e-96c5-195ee398cc7b {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 966.885833] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-599b8aff-df54-40bd-8f70-90e466876e68 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.892118] env[62385]: DEBUG oslo_vmware.api [None req-89a795a6-938d-4015-8160-ceaccf5f0691 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 966.892118] env[62385]: value = "task-1206213" [ 966.892118] env[62385]: _type = "Task" [ 966.892118] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.901780] env[62385]: DEBUG oslo_vmware.api [None req-89a795a6-938d-4015-8160-ceaccf5f0691 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206213, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.019920] env[62385]: DEBUG nova.network.neutron [req-f1419ec6-f4b5-4a84-9ac8-604846a54cfb req-a85029a7-f506-4a96-9e88-cb526c8f10d7 service nova] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 967.177971] env[62385]: DEBUG nova.network.neutron [req-f1419ec6-f4b5-4a84-9ac8-604846a54cfb req-a85029a7-f506-4a96-9e88-cb526c8f10d7 service nova] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.225755] env[62385]: DEBUG oslo_vmware.api [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206210, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.289645] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.290s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.290338] env[62385]: DEBUG nova.compute.manager [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 967.293399] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.510s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.295125] env[62385]: INFO nova.compute.claims [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 967.334918] env[62385]: DEBUG oslo_vmware.api [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206211, 'name': PowerOffVM_Task, 'duration_secs': 0.160439} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.335901] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 967.336150] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 967.336956] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c499a84c-74f8-4a21-9a50-19668d7a9031 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.344824] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 967.345092] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-11e3aa9d-24ad-4a50-9397-bf0d010b9400 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.371304] env[62385]: DEBUG oslo_vmware.api [None req-a0e27224-d55b-467b-b751-b4feec4d5360 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206202, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.403173] env[62385]: DEBUG oslo_vmware.api [None req-89a795a6-938d-4015-8160-ceaccf5f0691 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206213, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.131964} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.403445] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-89a795a6-938d-4015-8160-ceaccf5f0691 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 967.403630] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-89a795a6-938d-4015-8160-ceaccf5f0691 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 967.403864] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-89a795a6-938d-4015-8160-ceaccf5f0691 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 967.404008] env[62385]: INFO nova.compute.manager [None req-89a795a6-938d-4015-8160-ceaccf5f0691 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Took 1.12 seconds to destroy the instance on the hypervisor. [ 967.404269] env[62385]: DEBUG oslo.service.loopingcall [None req-89a795a6-938d-4015-8160-ceaccf5f0691 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 967.404463] env[62385]: DEBUG nova.compute.manager [-] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 967.404562] env[62385]: DEBUG nova.network.neutron [-] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 967.415693] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 967.415993] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 967.416281] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Deleting the datastore file [datastore1] fd0b59c2-e38d-4ecc-b32f-72f07d555ffd {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 967.416643] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fa4a722e-6630-4d73-97e7-60898053d02c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.423483] env[62385]: DEBUG oslo_vmware.api [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 967.423483] env[62385]: value = "task-1206215" [ 967.423483] env[62385]: _type = "Task" [ 967.423483] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.432133] env[62385]: DEBUG oslo_vmware.api [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206215, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.680987] env[62385]: DEBUG oslo_concurrency.lockutils [req-f1419ec6-f4b5-4a84-9ac8-604846a54cfb req-a85029a7-f506-4a96-9e88-cb526c8f10d7 service nova] Releasing lock "refresh_cache-b1e0baf3-643c-49c2-8a80-8de07d8527e7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.681362] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquired lock "refresh_cache-b1e0baf3-643c-49c2-8a80-8de07d8527e7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.681538] env[62385]: DEBUG nova.network.neutron [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 967.727056] env[62385]: DEBUG oslo_vmware.api [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206210, 'name': PowerOnVM_Task, 'duration_secs': 0.588186} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.727498] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 967.727781] env[62385]: INFO nova.compute.manager [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Took 8.59 seconds to spawn the instance on the hypervisor. [ 967.728057] env[62385]: DEBUG nova.compute.manager [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 967.729035] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63f081cc-6740-43d5-b76f-bbb8483add65 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.799692] env[62385]: DEBUG nova.compute.utils [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 967.805413] env[62385]: DEBUG nova.compute.manager [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 967.805413] env[62385]: DEBUG nova.network.neutron [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 967.873910] env[62385]: DEBUG oslo_vmware.api [None req-a0e27224-d55b-467b-b751-b4feec4d5360 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206202, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.899995] env[62385]: DEBUG nova.policy [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '12f9521f9c9b4c6b9afc4973be32c5b7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '731b6700ce5b43369b40165545f04950', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 967.933559] env[62385]: DEBUG oslo_vmware.api [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206215, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.248209] env[62385]: INFO nova.compute.manager [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Took 21.85 seconds to build instance. [ 968.258032] env[62385]: DEBUG nova.network.neutron [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 968.305320] env[62385]: DEBUG nova.compute.manager [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 968.374243] env[62385]: DEBUG oslo_vmware.api [None req-a0e27224-d55b-467b-b751-b4feec4d5360 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206202, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.399802] env[62385]: DEBUG nova.network.neutron [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Successfully created port: a755329d-e122-40af-add4-88f11f88402c {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 968.440061] env[62385]: DEBUG oslo_vmware.api [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206215, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.626693} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.440061] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 968.440481] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 968.440775] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 968.509204] env[62385]: DEBUG nova.compute.manager [req-8ab337cb-40a7-4885-b951-456a495360af req-668cc78d-7116-4e35-bf00-57de349629d4 service nova] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Received event network-vif-deleted-77d05c5a-39f5-45b0-a17c-599e5a2c06fd {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 968.509204] env[62385]: INFO nova.compute.manager [req-8ab337cb-40a7-4885-b951-456a495360af req-668cc78d-7116-4e35-bf00-57de349629d4 service nova] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Neutron deleted interface 77d05c5a-39f5-45b0-a17c-599e5a2c06fd; detaching it from the instance and deleting it from the info cache [ 968.509204] env[62385]: DEBUG nova.network.neutron [req-8ab337cb-40a7-4885-b951-456a495360af req-668cc78d-7116-4e35-bf00-57de349629d4 service nova] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.541351] env[62385]: DEBUG nova.network.neutron [-] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.593952] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fed5e651-d089-42e3-89cc-a64f30fd2e16 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.602887] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8c19063-b105-47ad-bb4a-989b13a697d2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.636529] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b69b0da2-3ddc-4d21-842e-3690223c727b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.645507] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b06e7db4-4b4b-476b-a148-20c98e05b2d3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.650866] env[62385]: DEBUG nova.network.neutron [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Updating instance_info_cache with network_info: [{"id": "eef524f4-fc00-47f1-a485-dcf568c6d0e4", "address": "fa:16:3e:e6:2c:55", "network": {"id": "ff21dd88-fa02-4bd6-8009-980ec6e72758", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1065789169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aea48968d4464ac0ab01c998c059107d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3ff3baee-99ce-4b51-ae98-efc6163aaab3", "external-id": "nsx-vlan-transportzone-574", "segmentation_id": 574, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeef524f4-fc", "ovs_interfaceid": "eef524f4-fc00-47f1-a485-dcf568c6d0e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.663501] env[62385]: DEBUG nova.compute.provider_tree [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 968.750821] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bd34edc8-5faa-4e29-a0da-77c4297c1a2d tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "df020d6b-3fab-4599-a342-47c7833b4240" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.365s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.876543] env[62385]: DEBUG oslo_vmware.api [None req-a0e27224-d55b-467b-b751-b4feec4d5360 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206202, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.011030] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4b51f10b-419c-4e63-9cfa-8d896863175b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.020482] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74121993-59ac-45a9-a3bf-6f66483169a6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.051855] env[62385]: INFO nova.compute.manager [-] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Took 1.65 seconds to deallocate network for instance. [ 969.052632] env[62385]: DEBUG nova.compute.manager [req-8ab337cb-40a7-4885-b951-456a495360af req-668cc78d-7116-4e35-bf00-57de349629d4 service nova] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Detach interface failed, port_id=77d05c5a-39f5-45b0-a17c-599e5a2c06fd, reason: Instance 9b2f8292-9d89-407e-96c5-195ee398cc7b could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 969.154365] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Releasing lock "refresh_cache-b1e0baf3-643c-49c2-8a80-8de07d8527e7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.154687] env[62385]: DEBUG nova.compute.manager [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Instance network_info: |[{"id": "eef524f4-fc00-47f1-a485-dcf568c6d0e4", "address": "fa:16:3e:e6:2c:55", "network": {"id": "ff21dd88-fa02-4bd6-8009-980ec6e72758", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1065789169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aea48968d4464ac0ab01c998c059107d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3ff3baee-99ce-4b51-ae98-efc6163aaab3", "external-id": "nsx-vlan-transportzone-574", "segmentation_id": 574, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeef524f4-fc", "ovs_interfaceid": "eef524f4-fc00-47f1-a485-dcf568c6d0e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 969.155271] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e6:2c:55', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3ff3baee-99ce-4b51-ae98-efc6163aaab3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'eef524f4-fc00-47f1-a485-dcf568c6d0e4', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 969.165830] env[62385]: DEBUG oslo.service.loopingcall [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 969.166884] env[62385]: DEBUG nova.scheduler.client.report [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 969.172501] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 969.173094] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f6690062-6b71-4a9f-9f5e-8909355cf5b3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.193710] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 969.193710] env[62385]: value = "task-1206216" [ 969.193710] env[62385]: _type = "Task" [ 969.193710] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.202221] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206216, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.323622] env[62385]: DEBUG nova.compute.manager [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 969.351092] env[62385]: DEBUG nova.virt.hardware [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 969.351385] env[62385]: DEBUG nova.virt.hardware [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 969.351551] env[62385]: DEBUG nova.virt.hardware [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 969.351736] env[62385]: DEBUG nova.virt.hardware [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 969.351911] env[62385]: DEBUG nova.virt.hardware [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 969.352055] env[62385]: DEBUG nova.virt.hardware [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 969.352278] env[62385]: DEBUG nova.virt.hardware [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 969.352443] env[62385]: DEBUG nova.virt.hardware [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 969.352619] env[62385]: DEBUG nova.virt.hardware [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 969.352789] env[62385]: DEBUG nova.virt.hardware [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 969.352970] env[62385]: DEBUG nova.virt.hardware [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 969.353857] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5c2caa0-9286-4630-93fc-6fb6aa59ef4b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.368601] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ddba296-9981-49cf-9928-17de026b1900 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.380188] env[62385]: DEBUG oslo_vmware.api [None req-a0e27224-d55b-467b-b751-b4feec4d5360 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206202, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.477200] env[62385]: DEBUG nova.virt.hardware [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 969.477554] env[62385]: DEBUG nova.virt.hardware [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 969.478165] env[62385]: DEBUG nova.virt.hardware [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 969.478165] env[62385]: DEBUG nova.virt.hardware [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 969.478335] env[62385]: DEBUG nova.virt.hardware [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 969.478481] env[62385]: DEBUG nova.virt.hardware [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 969.478908] env[62385]: DEBUG nova.virt.hardware [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 969.479165] env[62385]: DEBUG nova.virt.hardware [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 969.479368] env[62385]: DEBUG nova.virt.hardware [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 969.479844] env[62385]: DEBUG nova.virt.hardware [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 969.479844] env[62385]: DEBUG nova.virt.hardware [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 969.480943] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3f478f9-6f85-4f7a-9b6f-30c9ea3b5228 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.489027] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a1812ec-ed28-4be9-8121-0c158bbbc8b7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.503437] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7d:b9:30', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '39cd75b0-9ec7-48ed-b57f-34da0c573a60', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '31a45e90-bf05-4452-8a2c-490aa557ac1b', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 969.510765] env[62385]: DEBUG oslo.service.loopingcall [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 969.511054] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 969.511275] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9504c7d7-5956-4300-92ce-82c50eff8b0d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.530467] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 969.530467] env[62385]: value = "task-1206217" [ 969.530467] env[62385]: _type = "Task" [ 969.530467] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.538294] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206217, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.561462] env[62385]: DEBUG oslo_concurrency.lockutils [None req-89a795a6-938d-4015-8160-ceaccf5f0691 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.675134] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.381s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.675913] env[62385]: DEBUG nova.compute.manager [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 969.680229] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9dfd66a2-f29e-47d2-9366-04a5ee79630a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.287s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.680584] env[62385]: DEBUG nova.objects.instance [None req-9dfd66a2-f29e-47d2-9366-04a5ee79630a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lazy-loading 'resources' on Instance uuid bd7263b2-b996-4794-946b-2c28215574cb {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 969.704932] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206216, 'name': CreateVM_Task, 'duration_secs': 0.299806} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.704932] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 969.705680] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.705938] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.706453] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 969.706566] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-09c85a00-a029-4a6d-bde6-6e2c322dd15b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.711863] env[62385]: DEBUG oslo_vmware.api [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 969.711863] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]529f3fd8-bbe7-f1a7-caf5-98b0df107083" [ 969.711863] env[62385]: _type = "Task" [ 969.711863] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.720187] env[62385]: DEBUG oslo_vmware.api [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]529f3fd8-bbe7-f1a7-caf5-98b0df107083, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.874781] env[62385]: DEBUG oslo_vmware.api [None req-a0e27224-d55b-467b-b751-b4feec4d5360 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206202, 'name': ReconfigVM_Task, 'duration_secs': 5.811059} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.875100] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a0e27224-d55b-467b-b751-b4feec4d5360 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Releasing lock "f937ef7b-ab39-4d8a-9577-7faabc652aa5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.875326] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a0e27224-d55b-467b-b751-b4feec4d5360 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Reconfigured VM to detach interface {{(pid=62385) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 970.041896] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206217, 'name': CreateVM_Task, 'duration_secs': 0.313898} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.042286] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 970.042624] env[62385]: DEBUG oslo_concurrency.lockutils [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.184588] env[62385]: DEBUG nova.compute.utils [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 970.186142] env[62385]: DEBUG nova.compute.manager [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 970.186230] env[62385]: DEBUG nova.network.neutron [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 970.228413] env[62385]: DEBUG oslo_vmware.api [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]529f3fd8-bbe7-f1a7-caf5-98b0df107083, 'name': SearchDatastore_Task, 'duration_secs': 0.009518} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.228861] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.229427] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 970.229519] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.229735] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.229943] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 970.230262] env[62385]: DEBUG oslo_concurrency.lockutils [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.230576] env[62385]: DEBUG oslo_concurrency.lockutils [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 970.230819] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e97ff9dc-c865-4443-852f-afb6777ed06b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.233153] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-90b2076e-5ff3-4535-8794-10777339add4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.244698] env[62385]: DEBUG oslo_vmware.api [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 970.244698] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5252cd0f-734f-50a2-8b4f-b0f27de036b9" [ 970.244698] env[62385]: _type = "Task" [ 970.244698] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.249654] env[62385]: DEBUG nova.policy [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8988632e3af942a884b9c811e4ca539c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6e99f1e4cfd84bf083c8261b2ffcfa2f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 970.252469] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 970.252684] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 970.256065] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-64e9440e-3573-4253-9533-42f19e5928ff {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.267766] env[62385]: DEBUG oslo_vmware.api [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5252cd0f-734f-50a2-8b4f-b0f27de036b9, 'name': SearchDatastore_Task, 'duration_secs': 0.009158} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.268716] env[62385]: DEBUG oslo_concurrency.lockutils [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.268997] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 970.269320] env[62385]: DEBUG oslo_concurrency.lockutils [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.274640] env[62385]: DEBUG oslo_vmware.api [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 970.274640] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5212fd39-b077-336a-ede3-6d50f21f8627" [ 970.274640] env[62385]: _type = "Task" [ 970.274640] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.284498] env[62385]: DEBUG oslo_vmware.api [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5212fd39-b077-336a-ede3-6d50f21f8627, 'name': SearchDatastore_Task, 'duration_secs': 0.009359} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.287875] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f681dee5-48c5-4719-be09-56ad46476032 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.294430] env[62385]: DEBUG oslo_vmware.api [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 970.294430] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]524c6aea-c08c-e395-fda2-562f341aa538" [ 970.294430] env[62385]: _type = "Task" [ 970.294430] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.304787] env[62385]: DEBUG oslo_vmware.api [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]524c6aea-c08c-e395-fda2-562f341aa538, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.357767] env[62385]: DEBUG nova.network.neutron [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Successfully updated port: a755329d-e122-40af-add4-88f11f88402c {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 970.447838] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42b83411-94a1-494b-a2f8-d0a0ca851740 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.455780] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-939239b3-f0da-4680-8389-143dc54f25eb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.488686] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e25b3804-4459-461d-bc2c-d9f98b3a85d7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.497052] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f770eb0d-2cd2-4873-abef-ebcab604207e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.510738] env[62385]: DEBUG nova.compute.provider_tree [None req-9dfd66a2-f29e-47d2-9366-04a5ee79630a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 970.558510] env[62385]: DEBUG nova.compute.manager [req-153ae9ef-20a6-467f-acc3-7f075e308af9 req-07d74982-dd97-496a-9fe9-bbccbb02d2d6 service nova] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Received event network-vif-plugged-a755329d-e122-40af-add4-88f11f88402c {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 970.558874] env[62385]: DEBUG oslo_concurrency.lockutils [req-153ae9ef-20a6-467f-acc3-7f075e308af9 req-07d74982-dd97-496a-9fe9-bbccbb02d2d6 service nova] Acquiring lock "9741e9c8-60d9-41bf-8185-ef3cfcc1c844-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.559246] env[62385]: DEBUG oslo_concurrency.lockutils [req-153ae9ef-20a6-467f-acc3-7f075e308af9 req-07d74982-dd97-496a-9fe9-bbccbb02d2d6 service nova] Lock "9741e9c8-60d9-41bf-8185-ef3cfcc1c844-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.559542] env[62385]: DEBUG oslo_concurrency.lockutils [req-153ae9ef-20a6-467f-acc3-7f075e308af9 req-07d74982-dd97-496a-9fe9-bbccbb02d2d6 service nova] Lock "9741e9c8-60d9-41bf-8185-ef3cfcc1c844-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.559850] env[62385]: DEBUG nova.compute.manager [req-153ae9ef-20a6-467f-acc3-7f075e308af9 req-07d74982-dd97-496a-9fe9-bbccbb02d2d6 service nova] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] No waiting events found dispatching network-vif-plugged-a755329d-e122-40af-add4-88f11f88402c {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 970.560166] env[62385]: WARNING nova.compute.manager [req-153ae9ef-20a6-467f-acc3-7f075e308af9 req-07d74982-dd97-496a-9fe9-bbccbb02d2d6 service nova] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Received unexpected event network-vif-plugged-a755329d-e122-40af-add4-88f11f88402c for instance with vm_state building and task_state spawning. [ 970.560452] env[62385]: DEBUG nova.compute.manager [req-153ae9ef-20a6-467f-acc3-7f075e308af9 req-07d74982-dd97-496a-9fe9-bbccbb02d2d6 service nova] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Received event network-vif-deleted-c053ce8b-fc9f-4c48-9073-cdd753630700 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 970.560722] env[62385]: INFO nova.compute.manager [req-153ae9ef-20a6-467f-acc3-7f075e308af9 req-07d74982-dd97-496a-9fe9-bbccbb02d2d6 service nova] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Neutron deleted interface c053ce8b-fc9f-4c48-9073-cdd753630700; detaching it from the instance and deleting it from the info cache [ 970.561179] env[62385]: DEBUG nova.network.neutron [req-153ae9ef-20a6-467f-acc3-7f075e308af9 req-07d74982-dd97-496a-9fe9-bbccbb02d2d6 service nova] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Updating instance_info_cache with network_info: [{"id": "2b05b2b6-d4ce-4eb7-9d65-730d4dbad51c", "address": "fa:16:3e:ec:d7:16", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b05b2b6-d4", "ovs_interfaceid": "2b05b2b6-d4ce-4eb7-9d65-730d4dbad51c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "091890dd-d345-4cee-8b38-f7f812092c5f", "address": "fa:16:3e:b3:d9:f5", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap091890dd-d3", "ovs_interfaceid": "091890dd-d345-4cee-8b38-f7f812092c5f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.589011] env[62385]: DEBUG nova.network.neutron [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Successfully created port: 20926145-a8ac-4f6f-afb1-a1c449d1d76d {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 970.659882] env[62385]: DEBUG nova.compute.manager [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Stashing vm_state: active {{(pid=62385) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 970.693015] env[62385]: DEBUG nova.compute.manager [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 970.805063] env[62385]: DEBUG oslo_vmware.api [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]524c6aea-c08c-e395-fda2-562f341aa538, 'name': SearchDatastore_Task, 'duration_secs': 0.009599} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.805361] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.805650] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] b1e0baf3-643c-49c2-8a80-8de07d8527e7/b1e0baf3-643c-49c2-8a80-8de07d8527e7.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 970.805942] env[62385]: DEBUG oslo_concurrency.lockutils [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.806163] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 970.806399] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ed5c2d5f-b05b-4f4c-8ffd-369563dc356f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.808669] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5c6538c2-2162-4b27-a372-4f69cb83495f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.816429] env[62385]: DEBUG oslo_vmware.api [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 970.816429] env[62385]: value = "task-1206218" [ 970.816429] env[62385]: _type = "Task" [ 970.816429] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.817513] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 970.817701] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 970.821134] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-48086816-1755-46f6-b1d9-c9f94d1102d3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.826799] env[62385]: DEBUG oslo_vmware.api [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 970.826799] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c8386b-5cd5-9396-6218-47db74df01d4" [ 970.826799] env[62385]: _type = "Task" [ 970.826799] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.830049] env[62385]: DEBUG oslo_vmware.api [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206218, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.838469] env[62385]: DEBUG oslo_vmware.api [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c8386b-5cd5-9396-6218-47db74df01d4, 'name': SearchDatastore_Task, 'duration_secs': 0.008506} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.839264] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-934bd03c-21ee-4fed-93a5-e7e3f03f62f4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.844400] env[62385]: DEBUG oslo_vmware.api [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 970.844400] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52578abd-85b2-8739-fcd0-cdfc24b69bc5" [ 970.844400] env[62385]: _type = "Task" [ 970.844400] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.852508] env[62385]: DEBUG oslo_vmware.api [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52578abd-85b2-8739-fcd0-cdfc24b69bc5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.860236] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquiring lock "refresh_cache-9741e9c8-60d9-41bf-8185-ef3cfcc1c844" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.860382] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquired lock "refresh_cache-9741e9c8-60d9-41bf-8185-ef3cfcc1c844" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.860516] env[62385]: DEBUG nova.network.neutron [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 971.017181] env[62385]: DEBUG nova.scheduler.client.report [None req-9dfd66a2-f29e-47d2-9366-04a5ee79630a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 971.068061] env[62385]: DEBUG oslo_concurrency.lockutils [req-153ae9ef-20a6-467f-acc3-7f075e308af9 req-07d74982-dd97-496a-9fe9-bbccbb02d2d6 service nova] Acquiring lock "f937ef7b-ab39-4d8a-9577-7faabc652aa5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.068061] env[62385]: DEBUG oslo_concurrency.lockutils [req-153ae9ef-20a6-467f-acc3-7f075e308af9 req-07d74982-dd97-496a-9fe9-bbccbb02d2d6 service nova] Acquired lock "f937ef7b-ab39-4d8a-9577-7faabc652aa5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.069586] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdeff5f0-7a21-49a6-aae5-44df0d1e85ff {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.090492] env[62385]: DEBUG oslo_concurrency.lockutils [req-153ae9ef-20a6-467f-acc3-7f075e308af9 req-07d74982-dd97-496a-9fe9-bbccbb02d2d6 service nova] Releasing lock "f937ef7b-ab39-4d8a-9577-7faabc652aa5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.090814] env[62385]: WARNING nova.compute.manager [req-153ae9ef-20a6-467f-acc3-7f075e308af9 req-07d74982-dd97-496a-9fe9-bbccbb02d2d6 service nova] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Detach interface failed, port_id=c053ce8b-fc9f-4c48-9073-cdd753630700, reason: No device with interface-id c053ce8b-fc9f-4c48-9073-cdd753630700 exists on VM: nova.exception.NotFound: No device with interface-id c053ce8b-fc9f-4c48-9073-cdd753630700 exists on VM [ 971.091039] env[62385]: DEBUG nova.compute.manager [req-153ae9ef-20a6-467f-acc3-7f075e308af9 req-07d74982-dd97-496a-9fe9-bbccbb02d2d6 service nova] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Received event network-changed-a755329d-e122-40af-add4-88f11f88402c {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 971.091222] env[62385]: DEBUG nova.compute.manager [req-153ae9ef-20a6-467f-acc3-7f075e308af9 req-07d74982-dd97-496a-9fe9-bbccbb02d2d6 service nova] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Refreshing instance network info cache due to event network-changed-a755329d-e122-40af-add4-88f11f88402c. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 971.091640] env[62385]: DEBUG oslo_concurrency.lockutils [req-153ae9ef-20a6-467f-acc3-7f075e308af9 req-07d74982-dd97-496a-9fe9-bbccbb02d2d6 service nova] Acquiring lock "refresh_cache-9741e9c8-60d9-41bf-8185-ef3cfcc1c844" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.188052] env[62385]: DEBUG oslo_concurrency.lockutils [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.271207] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a0e27224-d55b-467b-b751-b4feec4d5360 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "refresh_cache-f937ef7b-ab39-4d8a-9577-7faabc652aa5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.271456] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a0e27224-d55b-467b-b751-b4feec4d5360 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquired lock "refresh_cache-f937ef7b-ab39-4d8a-9577-7faabc652aa5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.271624] env[62385]: DEBUG nova.network.neutron [None req-a0e27224-d55b-467b-b751-b4feec4d5360 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 971.327066] env[62385]: DEBUG oslo_vmware.api [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206218, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.443186} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.327229] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] b1e0baf3-643c-49c2-8a80-8de07d8527e7/b1e0baf3-643c-49c2-8a80-8de07d8527e7.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 971.327354] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 971.327578] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-659eb199-dc78-4f7e-99cf-aeee50d08ff1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.333887] env[62385]: DEBUG oslo_vmware.api [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 971.333887] env[62385]: value = "task-1206219" [ 971.333887] env[62385]: _type = "Task" [ 971.333887] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.342889] env[62385]: DEBUG oslo_vmware.api [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206219, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.353577] env[62385]: DEBUG oslo_vmware.api [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52578abd-85b2-8739-fcd0-cdfc24b69bc5, 'name': SearchDatastore_Task, 'duration_secs': 0.008188} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.353860] env[62385]: DEBUG oslo_concurrency.lockutils [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.354159] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] fd0b59c2-e38d-4ecc-b32f-72f07d555ffd/fd0b59c2-e38d-4ecc-b32f-72f07d555ffd.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 971.354442] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dedc04b0-b6ec-44da-88ec-dd0231a33312 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.360547] env[62385]: DEBUG oslo_vmware.api [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 971.360547] env[62385]: value = "task-1206220" [ 971.360547] env[62385]: _type = "Task" [ 971.360547] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.371165] env[62385]: DEBUG oslo_vmware.api [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206220, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.394935] env[62385]: DEBUG nova.network.neutron [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 971.474577] env[62385]: DEBUG oslo_concurrency.lockutils [None req-74f8a615-f91b-4cab-a95b-682654db78d4 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "f937ef7b-ab39-4d8a-9577-7faabc652aa5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.474931] env[62385]: DEBUG oslo_concurrency.lockutils [None req-74f8a615-f91b-4cab-a95b-682654db78d4 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "f937ef7b-ab39-4d8a-9577-7faabc652aa5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.475204] env[62385]: DEBUG oslo_concurrency.lockutils [None req-74f8a615-f91b-4cab-a95b-682654db78d4 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "f937ef7b-ab39-4d8a-9577-7faabc652aa5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.475428] env[62385]: DEBUG oslo_concurrency.lockutils [None req-74f8a615-f91b-4cab-a95b-682654db78d4 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "f937ef7b-ab39-4d8a-9577-7faabc652aa5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.475623] env[62385]: DEBUG oslo_concurrency.lockutils [None req-74f8a615-f91b-4cab-a95b-682654db78d4 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "f937ef7b-ab39-4d8a-9577-7faabc652aa5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.478521] env[62385]: INFO nova.compute.manager [None req-74f8a615-f91b-4cab-a95b-682654db78d4 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Terminating instance [ 971.480751] env[62385]: DEBUG nova.compute.manager [None req-74f8a615-f91b-4cab-a95b-682654db78d4 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 971.481117] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-74f8a615-f91b-4cab-a95b-682654db78d4 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 971.482039] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81d69273-e7b1-4f51-a187-5a92456570f0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.494359] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-74f8a615-f91b-4cab-a95b-682654db78d4 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 971.494642] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4d80dfb4-6387-4667-b4c7-b110f48f17af {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.502117] env[62385]: DEBUG oslo_vmware.api [None req-74f8a615-f91b-4cab-a95b-682654db78d4 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 971.502117] env[62385]: value = "task-1206221" [ 971.502117] env[62385]: _type = "Task" [ 971.502117] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.513725] env[62385]: DEBUG oslo_vmware.api [None req-74f8a615-f91b-4cab-a95b-682654db78d4 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206221, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.522162] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9dfd66a2-f29e-47d2-9366-04a5ee79630a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.842s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.524970] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f6add1d4-a93d-4ad5-bbb2-9f5005aeead7 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.258s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.525303] env[62385]: DEBUG nova.objects.instance [None req-f6add1d4-a93d-4ad5-bbb2-9f5005aeead7 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lazy-loading 'resources' on Instance uuid 33387505-c576-488b-8c9c-b064fe81a7d7 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 971.562071] env[62385]: INFO nova.scheduler.client.report [None req-9dfd66a2-f29e-47d2-9366-04a5ee79630a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Deleted allocations for instance bd7263b2-b996-4794-946b-2c28215574cb [ 971.673988] env[62385]: DEBUG nova.network.neutron [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Updating instance_info_cache with network_info: [{"id": "a755329d-e122-40af-add4-88f11f88402c", "address": "fa:16:3e:da:40:d7", "network": {"id": "2c1e2031-ab59-47f5-a758-3a314aca6c1e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1100903482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "731b6700ce5b43369b40165545f04950", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1316f5aa-529f-4bac-8dd7-6076a9d43312", "external-id": "nsx-vlan-transportzone-399", "segmentation_id": 399, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa755329d-e1", "ovs_interfaceid": "a755329d-e122-40af-add4-88f11f88402c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.708177] env[62385]: DEBUG nova.compute.manager [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 971.741816] env[62385]: DEBUG nova.virt.hardware [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 971.742105] env[62385]: DEBUG nova.virt.hardware [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 971.742283] env[62385]: DEBUG nova.virt.hardware [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 971.742469] env[62385]: DEBUG nova.virt.hardware [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 971.742621] env[62385]: DEBUG nova.virt.hardware [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 971.742772] env[62385]: DEBUG nova.virt.hardware [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 971.742990] env[62385]: DEBUG nova.virt.hardware [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 971.743176] env[62385]: DEBUG nova.virt.hardware [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 971.743352] env[62385]: DEBUG nova.virt.hardware [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 971.743519] env[62385]: DEBUG nova.virt.hardware [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 971.743698] env[62385]: DEBUG nova.virt.hardware [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 971.744633] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e171ffc8-7c94-404b-832d-eead213b6242 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.753518] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7df01de-bd65-488e-854f-7bd5f90517ab {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.849019] env[62385]: DEBUG oslo_vmware.api [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206219, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06269} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.849019] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 971.849019] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-175e586c-df63-48f8-9b45-4d58652e0364 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.872996] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] b1e0baf3-643c-49c2-8a80-8de07d8527e7/b1e0baf3-643c-49c2-8a80-8de07d8527e7.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 971.876105] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-42b91fe5-cfbe-430a-813d-a01b2a3b9fff {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.898491] env[62385]: DEBUG oslo_vmware.api [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206220, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.436623} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.899825] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] fd0b59c2-e38d-4ecc-b32f-72f07d555ffd/fd0b59c2-e38d-4ecc-b32f-72f07d555ffd.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 971.900088] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 971.900408] env[62385]: DEBUG oslo_vmware.api [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 971.900408] env[62385]: value = "task-1206222" [ 971.900408] env[62385]: _type = "Task" [ 971.900408] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.900655] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-814b6dee-7a6d-4923-8c7b-fbc5e90984c2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.911472] env[62385]: DEBUG oslo_vmware.api [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206222, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.912707] env[62385]: DEBUG oslo_vmware.api [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 971.912707] env[62385]: value = "task-1206223" [ 971.912707] env[62385]: _type = "Task" [ 971.912707] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.921187] env[62385]: DEBUG oslo_vmware.api [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206223, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.014093] env[62385]: DEBUG oslo_vmware.api [None req-74f8a615-f91b-4cab-a95b-682654db78d4 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206221, 'name': PowerOffVM_Task, 'duration_secs': 0.26268} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.015114] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-74f8a615-f91b-4cab-a95b-682654db78d4 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 972.015309] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-74f8a615-f91b-4cab-a95b-682654db78d4 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 972.015566] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-de9b80c6-3412-4fbf-bb6b-b6dd9aaf4d12 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.052884] env[62385]: INFO nova.network.neutron [None req-a0e27224-d55b-467b-b751-b4feec4d5360 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Port 091890dd-d345-4cee-8b38-f7f812092c5f from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 972.053268] env[62385]: DEBUG nova.network.neutron [None req-a0e27224-d55b-467b-b751-b4feec4d5360 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Updating instance_info_cache with network_info: [{"id": "2b05b2b6-d4ce-4eb7-9d65-730d4dbad51c", "address": "fa:16:3e:ec:d7:16", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b05b2b6-d4", "ovs_interfaceid": "2b05b2b6-d4ce-4eb7-9d65-730d4dbad51c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 972.073919] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9dfd66a2-f29e-47d2-9366-04a5ee79630a tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "bd7263b2-b996-4794-946b-2c28215574cb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.604s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.153337] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-74f8a615-f91b-4cab-a95b-682654db78d4 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 972.153569] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-74f8a615-f91b-4cab-a95b-682654db78d4 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 972.153755] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-74f8a615-f91b-4cab-a95b-682654db78d4 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Deleting the datastore file [datastore1] f937ef7b-ab39-4d8a-9577-7faabc652aa5 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 972.154078] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3e0c6de7-fe9f-406e-8713-eef672eb56b2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.160914] env[62385]: DEBUG oslo_vmware.api [None req-74f8a615-f91b-4cab-a95b-682654db78d4 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 972.160914] env[62385]: value = "task-1206225" [ 972.160914] env[62385]: _type = "Task" [ 972.160914] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.178086] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Releasing lock "refresh_cache-9741e9c8-60d9-41bf-8185-ef3cfcc1c844" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.178294] env[62385]: DEBUG nova.compute.manager [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Instance network_info: |[{"id": "a755329d-e122-40af-add4-88f11f88402c", "address": "fa:16:3e:da:40:d7", "network": {"id": "2c1e2031-ab59-47f5-a758-3a314aca6c1e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1100903482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "731b6700ce5b43369b40165545f04950", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1316f5aa-529f-4bac-8dd7-6076a9d43312", "external-id": "nsx-vlan-transportzone-399", "segmentation_id": 399, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa755329d-e1", "ovs_interfaceid": "a755329d-e122-40af-add4-88f11f88402c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 972.178748] env[62385]: DEBUG oslo_vmware.api [None req-74f8a615-f91b-4cab-a95b-682654db78d4 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206225, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.181997] env[62385]: DEBUG oslo_concurrency.lockutils [req-153ae9ef-20a6-467f-acc3-7f075e308af9 req-07d74982-dd97-496a-9fe9-bbccbb02d2d6 service nova] Acquired lock "refresh_cache-9741e9c8-60d9-41bf-8185-ef3cfcc1c844" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.182224] env[62385]: DEBUG nova.network.neutron [req-153ae9ef-20a6-467f-acc3-7f075e308af9 req-07d74982-dd97-496a-9fe9-bbccbb02d2d6 service nova] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Refreshing network info cache for port a755329d-e122-40af-add4-88f11f88402c {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 972.184399] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:da:40:d7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1316f5aa-529f-4bac-8dd7-6076a9d43312', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a755329d-e122-40af-add4-88f11f88402c', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 972.193362] env[62385]: DEBUG oslo.service.loopingcall [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 972.195260] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 972.196673] env[62385]: DEBUG nova.compute.manager [req-2d89e626-494b-45aa-8ac6-1a896b9bfd20 req-51358f36-e3b5-40b1-a66d-b76580036000 service nova] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Received event network-vif-plugged-20926145-a8ac-4f6f-afb1-a1c449d1d76d {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 972.196873] env[62385]: DEBUG oslo_concurrency.lockutils [req-2d89e626-494b-45aa-8ac6-1a896b9bfd20 req-51358f36-e3b5-40b1-a66d-b76580036000 service nova] Acquiring lock "a1c71d8b-0953-4cc2-961d-0c2b2fc11110-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.197102] env[62385]: DEBUG oslo_concurrency.lockutils [req-2d89e626-494b-45aa-8ac6-1a896b9bfd20 req-51358f36-e3b5-40b1-a66d-b76580036000 service nova] Lock "a1c71d8b-0953-4cc2-961d-0c2b2fc11110-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.197276] env[62385]: DEBUG oslo_concurrency.lockutils [req-2d89e626-494b-45aa-8ac6-1a896b9bfd20 req-51358f36-e3b5-40b1-a66d-b76580036000 service nova] Lock "a1c71d8b-0953-4cc2-961d-0c2b2fc11110-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.197447] env[62385]: DEBUG nova.compute.manager [req-2d89e626-494b-45aa-8ac6-1a896b9bfd20 req-51358f36-e3b5-40b1-a66d-b76580036000 service nova] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] No waiting events found dispatching network-vif-plugged-20926145-a8ac-4f6f-afb1-a1c449d1d76d {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 972.197621] env[62385]: WARNING nova.compute.manager [req-2d89e626-494b-45aa-8ac6-1a896b9bfd20 req-51358f36-e3b5-40b1-a66d-b76580036000 service nova] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Received unexpected event network-vif-plugged-20926145-a8ac-4f6f-afb1-a1c449d1d76d for instance with vm_state building and task_state spawning. [ 972.198410] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-63b982ad-5920-4a79-8956-d03c02071c15 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.223456] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 972.223456] env[62385]: value = "task-1206226" [ 972.223456] env[62385]: _type = "Task" [ 972.223456] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.239026] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206226, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.277493] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b0a4cbf-4781-46f3-a26a-f288aa7b1e08 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.285939] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c9ad407-e153-447e-9b30-08dc97ffaff9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.322522] env[62385]: DEBUG nova.network.neutron [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Successfully updated port: 20926145-a8ac-4f6f-afb1-a1c449d1d76d {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 972.324562] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb1233a8-a64b-422e-bdb5-1aa0e0626f77 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.333749] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc2b864e-92e7-428a-95fc-e0e63760f852 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.349054] env[62385]: DEBUG nova.compute.provider_tree [None req-f6add1d4-a93d-4ad5-bbb2-9f5005aeead7 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 972.412821] env[62385]: DEBUG oslo_vmware.api [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206222, 'name': ReconfigVM_Task, 'duration_secs': 0.288116} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.413223] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Reconfigured VM instance instance-00000061 to attach disk [datastore2] b1e0baf3-643c-49c2-8a80-8de07d8527e7/b1e0baf3-643c-49c2-8a80-8de07d8527e7.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 972.417181] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1c849901-762d-4005-b338-4a13f735a760 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.426373] env[62385]: DEBUG oslo_vmware.api [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206223, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.220343} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.426685] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 972.428047] env[62385]: DEBUG oslo_vmware.api [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 972.428047] env[62385]: value = "task-1206227" [ 972.428047] env[62385]: _type = "Task" [ 972.428047] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.428047] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82649e1e-e3b8-4e7e-b781-941f48b2873d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.438578] env[62385]: DEBUG oslo_vmware.api [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206227, 'name': Rename_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.456928] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] fd0b59c2-e38d-4ecc-b32f-72f07d555ffd/fd0b59c2-e38d-4ecc-b32f-72f07d555ffd.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 972.457274] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ff37fde3-bfdc-4283-9c2a-63e1f2a98b7a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.477701] env[62385]: DEBUG oslo_vmware.api [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 972.477701] env[62385]: value = "task-1206228" [ 972.477701] env[62385]: _type = "Task" [ 972.477701] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.486117] env[62385]: DEBUG oslo_vmware.api [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206228, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.556356] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a0e27224-d55b-467b-b751-b4feec4d5360 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Releasing lock "refresh_cache-f937ef7b-ab39-4d8a-9577-7faabc652aa5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.587640] env[62385]: DEBUG nova.compute.manager [req-7d07c688-64ed-44e5-a490-653e012e8c66 req-addd4388-c9a6-49a4-96f3-659ace09f0c4 service nova] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Received event network-vif-deleted-091890dd-d345-4cee-8b38-f7f812092c5f {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 972.675628] env[62385]: DEBUG oslo_vmware.api [None req-74f8a615-f91b-4cab-a95b-682654db78d4 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206225, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.171518} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.675952] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-74f8a615-f91b-4cab-a95b-682654db78d4 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 972.676205] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-74f8a615-f91b-4cab-a95b-682654db78d4 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 972.676426] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-74f8a615-f91b-4cab-a95b-682654db78d4 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 972.676645] env[62385]: INFO nova.compute.manager [None req-74f8a615-f91b-4cab-a95b-682654db78d4 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Took 1.20 seconds to destroy the instance on the hypervisor. [ 972.676934] env[62385]: DEBUG oslo.service.loopingcall [None req-74f8a615-f91b-4cab-a95b-682654db78d4 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 972.677186] env[62385]: DEBUG nova.compute.manager [-] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 972.677318] env[62385]: DEBUG nova.network.neutron [-] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 972.735672] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206226, 'name': CreateVM_Task} progress is 25%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.752714] env[62385]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port 091890dd-d345-4cee-8b38-f7f812092c5f could not be found.", "detail": ""}} {{(pid=62385) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 972.752975] env[62385]: DEBUG nova.network.neutron [-] Unable to show port 091890dd-d345-4cee-8b38-f7f812092c5f as it no longer exists. {{(pid=62385) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 972.828545] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Acquiring lock "refresh_cache-a1c71d8b-0953-4cc2-961d-0c2b2fc11110" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.828686] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Acquired lock "refresh_cache-a1c71d8b-0953-4cc2-961d-0c2b2fc11110" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.828834] env[62385]: DEBUG nova.network.neutron [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 972.851820] env[62385]: DEBUG nova.scheduler.client.report [None req-f6add1d4-a93d-4ad5-bbb2-9f5005aeead7 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 972.940203] env[62385]: DEBUG oslo_vmware.api [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206227, 'name': Rename_Task, 'duration_secs': 0.190561} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.940563] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 972.940816] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-52994629-c14a-4f32-beaf-fc085d5102b5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.947119] env[62385]: DEBUG oslo_vmware.api [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 972.947119] env[62385]: value = "task-1206229" [ 972.947119] env[62385]: _type = "Task" [ 972.947119] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.955651] env[62385]: DEBUG oslo_vmware.api [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206229, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.988454] env[62385]: DEBUG oslo_vmware.api [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206228, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.019176] env[62385]: DEBUG nova.network.neutron [req-153ae9ef-20a6-467f-acc3-7f075e308af9 req-07d74982-dd97-496a-9fe9-bbccbb02d2d6 service nova] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Updated VIF entry in instance network info cache for port a755329d-e122-40af-add4-88f11f88402c. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 973.019544] env[62385]: DEBUG nova.network.neutron [req-153ae9ef-20a6-467f-acc3-7f075e308af9 req-07d74982-dd97-496a-9fe9-bbccbb02d2d6 service nova] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Updating instance_info_cache with network_info: [{"id": "a755329d-e122-40af-add4-88f11f88402c", "address": "fa:16:3e:da:40:d7", "network": {"id": "2c1e2031-ab59-47f5-a758-3a314aca6c1e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1100903482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "731b6700ce5b43369b40165545f04950", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1316f5aa-529f-4bac-8dd7-6076a9d43312", "external-id": "nsx-vlan-transportzone-399", "segmentation_id": 399, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa755329d-e1", "ovs_interfaceid": "a755329d-e122-40af-add4-88f11f88402c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.060885] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a0e27224-d55b-467b-b751-b4feec4d5360 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "interface-f937ef7b-ab39-4d8a-9577-7faabc652aa5-c053ce8b-fc9f-4c48-9073-cdd753630700" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.787s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.235690] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206226, 'name': CreateVM_Task} progress is 25%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.363508] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f6add1d4-a93d-4ad5-bbb2-9f5005aeead7 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.838s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.366329] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bebcac0a-2c03-43c8-925b-3f44500f4631 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.444s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.366789] env[62385]: DEBUG nova.objects.instance [None req-bebcac0a-2c03-43c8-925b-3f44500f4631 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Lazy-loading 'resources' on Instance uuid 104dea2c-6977-4683-bc32-05131858f562 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 973.372022] env[62385]: DEBUG nova.network.neutron [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 973.397862] env[62385]: INFO nova.scheduler.client.report [None req-f6add1d4-a93d-4ad5-bbb2-9f5005aeead7 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Deleted allocations for instance 33387505-c576-488b-8c9c-b064fe81a7d7 [ 973.462082] env[62385]: DEBUG oslo_vmware.api [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206229, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.488297] env[62385]: DEBUG oslo_vmware.api [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206228, 'name': ReconfigVM_Task, 'duration_secs': 0.811425} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.488582] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Reconfigured VM instance instance-0000005f to attach disk [datastore2] fd0b59c2-e38d-4ecc-b32f-72f07d555ffd/fd0b59c2-e38d-4ecc-b32f-72f07d555ffd.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 973.489291] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e48fc2bc-0922-4e8f-ad0a-e2a159d6b466 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.497481] env[62385]: DEBUG oslo_vmware.api [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 973.497481] env[62385]: value = "task-1206230" [ 973.497481] env[62385]: _type = "Task" [ 973.497481] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.506393] env[62385]: DEBUG oslo_vmware.api [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206230, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.522704] env[62385]: DEBUG oslo_concurrency.lockutils [req-153ae9ef-20a6-467f-acc3-7f075e308af9 req-07d74982-dd97-496a-9fe9-bbccbb02d2d6 service nova] Releasing lock "refresh_cache-9741e9c8-60d9-41bf-8185-ef3cfcc1c844" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.548751] env[62385]: DEBUG nova.network.neutron [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Updating instance_info_cache with network_info: [{"id": "20926145-a8ac-4f6f-afb1-a1c449d1d76d", "address": "fa:16:3e:98:24:99", "network": {"id": "aca904e1-fbd6-4def-bbec-bd70c0fc8af9", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-249813742-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e99f1e4cfd84bf083c8261b2ffcfa2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f7abe5d-b8fe-4983-bd50-e7469f1fe7f3", "external-id": "nsx-vlan-transportzone-263", "segmentation_id": 263, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap20926145-a8", "ovs_interfaceid": "20926145-a8ac-4f6f-afb1-a1c449d1d76d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.661488] env[62385]: DEBUG nova.network.neutron [-] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.735220] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206226, 'name': CreateVM_Task, 'duration_secs': 1.401377} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.735451] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 973.736173] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.736400] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.737031] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 973.737031] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d8747714-7685-4377-9405-39b0987e1a01 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.742436] env[62385]: DEBUG oslo_vmware.api [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 973.742436] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]526ce580-6676-b126-e66a-72f6b99c2970" [ 973.742436] env[62385]: _type = "Task" [ 973.742436] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.751165] env[62385]: DEBUG oslo_vmware.api [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]526ce580-6676-b126-e66a-72f6b99c2970, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.906778] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f6add1d4-a93d-4ad5-bbb2-9f5005aeead7 tempest-ServersAdminTestJSON-1487936049 tempest-ServersAdminTestJSON-1487936049-project-member] Lock "33387505-c576-488b-8c9c-b064fe81a7d7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.064s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.959000] env[62385]: DEBUG oslo_vmware.api [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206229, 'name': PowerOnVM_Task, 'duration_secs': 0.550905} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.959194] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 973.959397] env[62385]: INFO nova.compute.manager [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Took 8.80 seconds to spawn the instance on the hypervisor. [ 973.959580] env[62385]: DEBUG nova.compute.manager [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 973.960439] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86ed9447-9f03-454c-b983-4a649a50a94d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.011533] env[62385]: DEBUG oslo_vmware.api [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206230, 'name': Rename_Task, 'duration_secs': 0.15279} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.011811] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 974.012101] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-65e335eb-1ead-45f8-81f7-7182a3e68fac {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.018730] env[62385]: DEBUG oslo_vmware.api [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 974.018730] env[62385]: value = "task-1206231" [ 974.018730] env[62385]: _type = "Task" [ 974.018730] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.026647] env[62385]: DEBUG oslo_vmware.api [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206231, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.053482] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Releasing lock "refresh_cache-a1c71d8b-0953-4cc2-961d-0c2b2fc11110" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.053779] env[62385]: DEBUG nova.compute.manager [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Instance network_info: |[{"id": "20926145-a8ac-4f6f-afb1-a1c449d1d76d", "address": "fa:16:3e:98:24:99", "network": {"id": "aca904e1-fbd6-4def-bbec-bd70c0fc8af9", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-249813742-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e99f1e4cfd84bf083c8261b2ffcfa2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f7abe5d-b8fe-4983-bd50-e7469f1fe7f3", "external-id": "nsx-vlan-transportzone-263", "segmentation_id": 263, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap20926145-a8", "ovs_interfaceid": "20926145-a8ac-4f6f-afb1-a1c449d1d76d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 974.054378] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:98:24:99', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2f7abe5d-b8fe-4983-bd50-e7469f1fe7f3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '20926145-a8ac-4f6f-afb1-a1c449d1d76d', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 974.062088] env[62385]: DEBUG oslo.service.loopingcall [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 974.062313] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 974.062592] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-68b3afb1-c16e-4a08-8cb0-dfc192f8786b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.080532] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36861c58-4503-40f4-8ea4-ccf677012b4b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.089014] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a76ca706-a27f-49ff-95c4-928c18cadfab {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.092329] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 974.092329] env[62385]: value = "task-1206232" [ 974.092329] env[62385]: _type = "Task" [ 974.092329] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.124032] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-647657d0-2e35-4b41-99b5-2cfda4b51866 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.126858] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206232, 'name': CreateVM_Task} progress is 15%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.131802] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cce19715-70e8-4701-8a19-d47d07274b79 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.147448] env[62385]: DEBUG nova.compute.provider_tree [None req-bebcac0a-2c03-43c8-925b-3f44500f4631 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 974.165147] env[62385]: INFO nova.compute.manager [-] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Took 1.49 seconds to deallocate network for instance. [ 974.229276] env[62385]: DEBUG nova.compute.manager [req-e0ebe632-a55e-4a70-a8a8-9516fee5f5eb req-eb9b9888-ecfc-425c-bfbd-085e33b9ef99 service nova] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Received event network-changed-20926145-a8ac-4f6f-afb1-a1c449d1d76d {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 974.229626] env[62385]: DEBUG nova.compute.manager [req-e0ebe632-a55e-4a70-a8a8-9516fee5f5eb req-eb9b9888-ecfc-425c-bfbd-085e33b9ef99 service nova] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Refreshing instance network info cache due to event network-changed-20926145-a8ac-4f6f-afb1-a1c449d1d76d. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 974.229977] env[62385]: DEBUG oslo_concurrency.lockutils [req-e0ebe632-a55e-4a70-a8a8-9516fee5f5eb req-eb9b9888-ecfc-425c-bfbd-085e33b9ef99 service nova] Acquiring lock "refresh_cache-a1c71d8b-0953-4cc2-961d-0c2b2fc11110" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.230801] env[62385]: DEBUG oslo_concurrency.lockutils [req-e0ebe632-a55e-4a70-a8a8-9516fee5f5eb req-eb9b9888-ecfc-425c-bfbd-085e33b9ef99 service nova] Acquired lock "refresh_cache-a1c71d8b-0953-4cc2-961d-0c2b2fc11110" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.231113] env[62385]: DEBUG nova.network.neutron [req-e0ebe632-a55e-4a70-a8a8-9516fee5f5eb req-eb9b9888-ecfc-425c-bfbd-085e33b9ef99 service nova] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Refreshing network info cache for port 20926145-a8ac-4f6f-afb1-a1c449d1d76d {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 974.253842] env[62385]: DEBUG oslo_vmware.api [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]526ce580-6676-b126-e66a-72f6b99c2970, 'name': SearchDatastore_Task, 'duration_secs': 0.009536} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.254176] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.254417] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 974.254654] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.254807] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.254986] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 974.255274] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-664d416e-bea7-46d7-9686-3262f017cc20 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.263713] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 974.263907] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 974.264657] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c3239391-45a8-4041-838a-a28947b06bab {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.270088] env[62385]: DEBUG oslo_vmware.api [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 974.270088] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ba90eb-6255-a65c-9f07-3478ac2a3f0e" [ 974.270088] env[62385]: _type = "Task" [ 974.270088] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.278178] env[62385]: DEBUG oslo_vmware.api [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ba90eb-6255-a65c-9f07-3478ac2a3f0e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.485812] env[62385]: INFO nova.compute.manager [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Took 23.50 seconds to build instance. [ 974.532964] env[62385]: DEBUG oslo_vmware.api [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206231, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.604118] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206232, 'name': CreateVM_Task, 'duration_secs': 0.334351} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.604315] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 974.605020] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.605193] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.605517] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 974.605783] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0408d7a0-6e7a-4203-a66c-59e57504e5dc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.614660] env[62385]: DEBUG oslo_vmware.api [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for the task: (returnval){ [ 974.614660] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ae1edf-f81b-c923-1bda-abb1651e2b2d" [ 974.614660] env[62385]: _type = "Task" [ 974.614660] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.624210] env[62385]: DEBUG oslo_vmware.api [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ae1edf-f81b-c923-1bda-abb1651e2b2d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.649515] env[62385]: DEBUG nova.scheduler.client.report [None req-bebcac0a-2c03-43c8-925b-3f44500f4631 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 974.672980] env[62385]: DEBUG oslo_concurrency.lockutils [None req-74f8a615-f91b-4cab-a95b-682654db78d4 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.781092] env[62385]: DEBUG oslo_vmware.api [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ba90eb-6255-a65c-9f07-3478ac2a3f0e, 'name': SearchDatastore_Task, 'duration_secs': 0.008931} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.783970] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d7d9db38-f9e6-4983-8eb8-fff4425840de {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.789686] env[62385]: DEBUG oslo_vmware.api [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 974.789686] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ddc82b-670e-db67-074c-1e0909644343" [ 974.789686] env[62385]: _type = "Task" [ 974.789686] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.797095] env[62385]: DEBUG oslo_vmware.api [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ddc82b-670e-db67-074c-1e0909644343, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.989380] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8fa2899f-3611-408b-b505-5fac602e3c92 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "b1e0baf3-643c-49c2-8a80-8de07d8527e7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.022s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.996221] env[62385]: DEBUG nova.network.neutron [req-e0ebe632-a55e-4a70-a8a8-9516fee5f5eb req-eb9b9888-ecfc-425c-bfbd-085e33b9ef99 service nova] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Updated VIF entry in instance network info cache for port 20926145-a8ac-4f6f-afb1-a1c449d1d76d. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 974.996586] env[62385]: DEBUG nova.network.neutron [req-e0ebe632-a55e-4a70-a8a8-9516fee5f5eb req-eb9b9888-ecfc-425c-bfbd-085e33b9ef99 service nova] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Updating instance_info_cache with network_info: [{"id": "20926145-a8ac-4f6f-afb1-a1c449d1d76d", "address": "fa:16:3e:98:24:99", "network": {"id": "aca904e1-fbd6-4def-bbec-bd70c0fc8af9", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-249813742-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e99f1e4cfd84bf083c8261b2ffcfa2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f7abe5d-b8fe-4983-bd50-e7469f1fe7f3", "external-id": "nsx-vlan-transportzone-263", "segmentation_id": 263, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap20926145-a8", "ovs_interfaceid": "20926145-a8ac-4f6f-afb1-a1c449d1d76d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 975.032529] env[62385]: DEBUG oslo_vmware.api [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206231, 'name': PowerOnVM_Task, 'duration_secs': 0.626443} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.032936] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 975.033020] env[62385]: DEBUG nova.compute.manager [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 975.033780] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-927b24c9-de78-4b65-8314-10bd14a924f4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.167010] env[62385]: DEBUG oslo_vmware.api [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ae1edf-f81b-c923-1bda-abb1651e2b2d, 'name': SearchDatastore_Task, 'duration_secs': 0.015273} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.167010] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.167010] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 975.167010] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.167010] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bebcac0a-2c03-43c8-925b-3f44500f4631 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.788s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.167010] env[62385]: DEBUG oslo_concurrency.lockutils [None req-89a795a6-938d-4015-8160-ceaccf5f0691 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.598s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.167010] env[62385]: DEBUG nova.objects.instance [None req-89a795a6-938d-4015-8160-ceaccf5f0691 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lazy-loading 'resources' on Instance uuid 9b2f8292-9d89-407e-96c5-195ee398cc7b {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 975.215202] env[62385]: INFO nova.scheduler.client.report [None req-bebcac0a-2c03-43c8-925b-3f44500f4631 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Deleted allocations for instance 104dea2c-6977-4683-bc32-05131858f562 [ 975.305023] env[62385]: DEBUG oslo_vmware.api [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ddc82b-670e-db67-074c-1e0909644343, 'name': SearchDatastore_Task, 'duration_secs': 0.009532} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.305023] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.305023] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 9741e9c8-60d9-41bf-8185-ef3cfcc1c844/9741e9c8-60d9-41bf-8185-ef3cfcc1c844.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 975.305023] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.305023] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 975.305023] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bac9bbe3-010a-4f16-b4fb-6cb13214e058 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.307735] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3ac4db78-1218-4220-ba6c-5964a27c3d2f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.318115] env[62385]: DEBUG oslo_vmware.api [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 975.318115] env[62385]: value = "task-1206233" [ 975.318115] env[62385]: _type = "Task" [ 975.318115] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.321715] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 975.323666] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 975.325326] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cfcdc10b-11ac-43c6-b943-371440b17edd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.333214] env[62385]: DEBUG oslo_vmware.api [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206233, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.338515] env[62385]: DEBUG oslo_vmware.api [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for the task: (returnval){ [ 975.338515] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5279e5a4-9aaa-8708-1595-4e26aa66c3c5" [ 975.338515] env[62385]: _type = "Task" [ 975.338515] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.350847] env[62385]: DEBUG oslo_vmware.api [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5279e5a4-9aaa-8708-1595-4e26aa66c3c5, 'name': SearchDatastore_Task, 'duration_secs': 0.009128} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.351896] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c2aabf2-6852-4264-a268-143d8a3e6454 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.358390] env[62385]: DEBUG oslo_vmware.api [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for the task: (returnval){ [ 975.358390] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5287ae4d-f1e4-9f7e-0014-211e4c0541f8" [ 975.358390] env[62385]: _type = "Task" [ 975.358390] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.366444] env[62385]: DEBUG oslo_vmware.api [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5287ae4d-f1e4-9f7e-0014-211e4c0541f8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.499223] env[62385]: DEBUG oslo_concurrency.lockutils [req-e0ebe632-a55e-4a70-a8a8-9516fee5f5eb req-eb9b9888-ecfc-425c-bfbd-085e33b9ef99 service nova] Releasing lock "refresh_cache-a1c71d8b-0953-4cc2-961d-0c2b2fc11110" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.499578] env[62385]: DEBUG nova.compute.manager [req-e0ebe632-a55e-4a70-a8a8-9516fee5f5eb req-eb9b9888-ecfc-425c-bfbd-085e33b9ef99 service nova] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Received event network-vif-deleted-2b05b2b6-d4ce-4eb7-9d65-730d4dbad51c {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 975.555128] env[62385]: DEBUG oslo_concurrency.lockutils [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.724597] env[62385]: DEBUG oslo_concurrency.lockutils [None req-bebcac0a-2c03-43c8-925b-3f44500f4631 tempest-InstanceActionsV221TestJSON-972957912 tempest-InstanceActionsV221TestJSON-972957912-project-member] Lock "104dea2c-6977-4683-bc32-05131858f562" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.957s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.833912] env[62385]: DEBUG oslo_vmware.api [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206233, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.462511} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.834260] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 9741e9c8-60d9-41bf-8185-ef3cfcc1c844/9741e9c8-60d9-41bf-8185-ef3cfcc1c844.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 975.834538] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 975.836835] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2f214fa3-39be-4b2b-a1a7-a3a8b5c9e4ba {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.850274] env[62385]: DEBUG oslo_vmware.api [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 975.850274] env[62385]: value = "task-1206234" [ 975.850274] env[62385]: _type = "Task" [ 975.850274] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.861666] env[62385]: DEBUG oslo_vmware.api [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206234, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.871510] env[62385]: DEBUG oslo_vmware.api [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5287ae4d-f1e4-9f7e-0014-211e4c0541f8, 'name': SearchDatastore_Task, 'duration_secs': 0.008412} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.874291] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.874578] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] a1c71d8b-0953-4cc2-961d-0c2b2fc11110/a1c71d8b-0953-4cc2-961d-0c2b2fc11110.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 975.875072] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1aba7ee5-f816-4233-8cb4-c9a8317d2815 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.882213] env[62385]: DEBUG oslo_vmware.api [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for the task: (returnval){ [ 975.882213] env[62385]: value = "task-1206235" [ 975.882213] env[62385]: _type = "Task" [ 975.882213] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.892321] env[62385]: DEBUG oslo_vmware.api [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206235, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.905400] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fefd6557-99b0-4b0c-a386-01de32d7cfdb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.914852] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d6b1871-d76a-4150-9069-1937444ebb2a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.948688] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e9dad91-a313-42c5-bd2a-3f565abb58f5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.963245] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3152c97b-518a-40ae-8d32-857f470fc6f2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.979384] env[62385]: DEBUG nova.compute.provider_tree [None req-89a795a6-938d-4015-8160-ceaccf5f0691 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 976.360448] env[62385]: DEBUG oslo_vmware.api [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206234, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071626} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.360815] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 976.361893] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37fa19b0-8e6a-4ade-8632-4c1d646c38a9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.393751] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] 9741e9c8-60d9-41bf-8185-ef3cfcc1c844/9741e9c8-60d9-41bf-8185-ef3cfcc1c844.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 976.394313] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3d139a99-8ca3-4c68-b561-d613bf57b9bd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.417978] env[62385]: DEBUG oslo_vmware.api [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206235, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.448697} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.419290] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] a1c71d8b-0953-4cc2-961d-0c2b2fc11110/a1c71d8b-0953-4cc2-961d-0c2b2fc11110.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 976.419520] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 976.419870] env[62385]: DEBUG oslo_vmware.api [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 976.419870] env[62385]: value = "task-1206236" [ 976.419870] env[62385]: _type = "Task" [ 976.419870] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.420084] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a30c0718-0e5d-4c3b-8037-a69c6c646f10 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.432862] env[62385]: DEBUG oslo_vmware.api [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206236, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.434641] env[62385]: DEBUG oslo_vmware.api [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for the task: (returnval){ [ 976.434641] env[62385]: value = "task-1206237" [ 976.434641] env[62385]: _type = "Task" [ 976.434641] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.445880] env[62385]: DEBUG oslo_vmware.api [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206237, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.461252] env[62385]: DEBUG oslo_concurrency.lockutils [None req-73f55945-68c9-4e01-b622-3cd3be847fe4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "fd0b59c2-e38d-4ecc-b32f-72f07d555ffd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.461574] env[62385]: DEBUG oslo_concurrency.lockutils [None req-73f55945-68c9-4e01-b622-3cd3be847fe4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "fd0b59c2-e38d-4ecc-b32f-72f07d555ffd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.461909] env[62385]: DEBUG oslo_concurrency.lockutils [None req-73f55945-68c9-4e01-b622-3cd3be847fe4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "fd0b59c2-e38d-4ecc-b32f-72f07d555ffd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.462372] env[62385]: DEBUG oslo_concurrency.lockutils [None req-73f55945-68c9-4e01-b622-3cd3be847fe4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "fd0b59c2-e38d-4ecc-b32f-72f07d555ffd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.462675] env[62385]: DEBUG oslo_concurrency.lockutils [None req-73f55945-68c9-4e01-b622-3cd3be847fe4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "fd0b59c2-e38d-4ecc-b32f-72f07d555ffd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.467586] env[62385]: INFO nova.compute.manager [None req-73f55945-68c9-4e01-b622-3cd3be847fe4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Terminating instance [ 976.469562] env[62385]: DEBUG nova.compute.manager [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Stashing vm_state: active {{(pid=62385) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 976.478023] env[62385]: DEBUG nova.compute.manager [None req-73f55945-68c9-4e01-b622-3cd3be847fe4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 976.478023] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-73f55945-68c9-4e01-b622-3cd3be847fe4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 976.478023] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05d60b6d-f333-4e7f-9f77-111d235ba12a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.482856] env[62385]: DEBUG nova.scheduler.client.report [None req-89a795a6-938d-4015-8160-ceaccf5f0691 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 976.495437] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-73f55945-68c9-4e01-b622-3cd3be847fe4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 976.495873] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4759294b-22cd-4e03-ab59-b5630498078d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.507485] env[62385]: DEBUG oslo_vmware.api [None req-73f55945-68c9-4e01-b622-3cd3be847fe4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 976.507485] env[62385]: value = "task-1206238" [ 976.507485] env[62385]: _type = "Task" [ 976.507485] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.523403] env[62385]: DEBUG oslo_vmware.api [None req-73f55945-68c9-4e01-b622-3cd3be847fe4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206238, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.933141] env[62385]: DEBUG oslo_vmware.api [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206236, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.944269] env[62385]: DEBUG oslo_vmware.api [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206237, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.191739} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.944535] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 976.945337] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2d1bb42-181f-4f36-a46a-30c1b3488e5e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.971468] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] a1c71d8b-0953-4cc2-961d-0c2b2fc11110/a1c71d8b-0953-4cc2-961d-0c2b2fc11110.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 976.971885] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7f4706be-5192-42f7-aefb-7dda11e982c0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.991571] env[62385]: DEBUG oslo_concurrency.lockutils [None req-89a795a6-938d-4015-8160-ceaccf5f0691 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.832s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.994272] env[62385]: DEBUG oslo_concurrency.lockutils [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 5.806s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.999409] env[62385]: DEBUG oslo_vmware.api [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for the task: (returnval){ [ 976.999409] env[62385]: value = "task-1206239" [ 976.999409] env[62385]: _type = "Task" [ 976.999409] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.009242] env[62385]: DEBUG oslo_vmware.api [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206239, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.010474] env[62385]: DEBUG oslo_concurrency.lockutils [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.017953] env[62385]: INFO nova.scheduler.client.report [None req-89a795a6-938d-4015-8160-ceaccf5f0691 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Deleted allocations for instance 9b2f8292-9d89-407e-96c5-195ee398cc7b [ 977.026979] env[62385]: DEBUG oslo_vmware.api [None req-73f55945-68c9-4e01-b622-3cd3be847fe4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206238, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.434594] env[62385]: DEBUG oslo_vmware.api [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206236, 'name': ReconfigVM_Task, 'duration_secs': 0.754164} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.434866] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Reconfigured VM instance instance-00000062 to attach disk [datastore2] 9741e9c8-60d9-41bf-8185-ef3cfcc1c844/9741e9c8-60d9-41bf-8185-ef3cfcc1c844.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 977.435437] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9bf2f866-5cdc-4905-8142-adc851d50267 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.441604] env[62385]: DEBUG oslo_vmware.api [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 977.441604] env[62385]: value = "task-1206240" [ 977.441604] env[62385]: _type = "Task" [ 977.441604] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.449819] env[62385]: DEBUG oslo_vmware.api [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206240, 'name': Rename_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.502902] env[62385]: INFO nova.compute.claims [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 977.521100] env[62385]: DEBUG oslo_vmware.api [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206239, 'name': ReconfigVM_Task, 'duration_secs': 0.289706} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.524249] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Reconfigured VM instance instance-00000063 to attach disk [datastore2] a1c71d8b-0953-4cc2-961d-0c2b2fc11110/a1c71d8b-0953-4cc2-961d-0c2b2fc11110.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 977.524955] env[62385]: DEBUG oslo_vmware.api [None req-73f55945-68c9-4e01-b622-3cd3be847fe4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206238, 'name': PowerOffVM_Task, 'duration_secs': 0.550684} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.525244] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ffe8494a-c47e-4626-8e3c-d5a83be8b0e2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.527071] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-73f55945-68c9-4e01-b622-3cd3be847fe4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 977.527300] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-73f55945-68c9-4e01-b622-3cd3be847fe4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 977.531669] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0b63707f-7fdd-46e4-9071-2d851a5c6cb3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.531669] env[62385]: DEBUG oslo_concurrency.lockutils [None req-89a795a6-938d-4015-8160-ceaccf5f0691 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "9b2f8292-9d89-407e-96c5-195ee398cc7b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.251s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.538049] env[62385]: DEBUG oslo_vmware.api [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for the task: (returnval){ [ 977.538049] env[62385]: value = "task-1206241" [ 977.538049] env[62385]: _type = "Task" [ 977.538049] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.556048] env[62385]: DEBUG oslo_vmware.api [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206241, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.604716] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-73f55945-68c9-4e01-b622-3cd3be847fe4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 977.605035] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-73f55945-68c9-4e01-b622-3cd3be847fe4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Deleting contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 977.605397] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-73f55945-68c9-4e01-b622-3cd3be847fe4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Deleting the datastore file [datastore2] fd0b59c2-e38d-4ecc-b32f-72f07d555ffd {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 977.605582] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d3c60df3-c5be-4ac3-b616-62c545b36f09 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.612014] env[62385]: DEBUG oslo_vmware.api [None req-73f55945-68c9-4e01-b622-3cd3be847fe4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 977.612014] env[62385]: value = "task-1206243" [ 977.612014] env[62385]: _type = "Task" [ 977.612014] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.619917] env[62385]: DEBUG oslo_vmware.api [None req-73f55945-68c9-4e01-b622-3cd3be847fe4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206243, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.956089] env[62385]: DEBUG oslo_vmware.api [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206240, 'name': Rename_Task, 'duration_secs': 0.139867} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.956397] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 977.956657] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8a81d674-645a-4757-aadb-01cf3fe7aba7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.963618] env[62385]: DEBUG oslo_vmware.api [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 977.963618] env[62385]: value = "task-1206244" [ 977.963618] env[62385]: _type = "Task" [ 977.963618] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.972850] env[62385]: DEBUG oslo_vmware.api [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206244, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.012019] env[62385]: INFO nova.compute.resource_tracker [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Updating resource usage from migration 183b505f-acf8-4ca5-bd5b-bbc88a8c7fd6 [ 978.057171] env[62385]: DEBUG oslo_vmware.api [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206241, 'name': Rename_Task, 'duration_secs': 0.262923} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.057818] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 978.057895] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6b47d5cd-9f7d-45b9-a888-cffb829beabf {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.066286] env[62385]: DEBUG oslo_vmware.api [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for the task: (returnval){ [ 978.066286] env[62385]: value = "task-1206245" [ 978.066286] env[62385]: _type = "Task" [ 978.066286] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.075742] env[62385]: DEBUG oslo_vmware.api [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206245, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.129708] env[62385]: DEBUG oslo_vmware.api [None req-73f55945-68c9-4e01-b622-3cd3be847fe4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206243, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.430932} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.130336] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-73f55945-68c9-4e01-b622-3cd3be847fe4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 978.130336] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-73f55945-68c9-4e01-b622-3cd3be847fe4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Deleted contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 978.130444] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-73f55945-68c9-4e01-b622-3cd3be847fe4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 978.131105] env[62385]: INFO nova.compute.manager [None req-73f55945-68c9-4e01-b622-3cd3be847fe4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Took 1.66 seconds to destroy the instance on the hypervisor. [ 978.131105] env[62385]: DEBUG oslo.service.loopingcall [None req-73f55945-68c9-4e01-b622-3cd3be847fe4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 978.131105] env[62385]: DEBUG nova.compute.manager [-] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 978.131251] env[62385]: DEBUG nova.network.neutron [-] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 978.260425] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84882a5e-b6ae-4d48-95ce-23acbecb5de4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.267345] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24b7439e-bec6-4972-875b-7551a864b82c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.305328] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bd04c7d-4cdc-4399-ab8e-6957ef3b92e2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.313805] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d77b4ffc-cd8d-4a3f-a3bb-f2480a705100 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.328602] env[62385]: DEBUG nova.compute.provider_tree [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 978.478505] env[62385]: DEBUG oslo_vmware.api [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206244, 'name': PowerOnVM_Task, 'duration_secs': 0.504188} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.479054] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 978.479361] env[62385]: INFO nova.compute.manager [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Took 9.16 seconds to spawn the instance on the hypervisor. [ 978.479694] env[62385]: DEBUG nova.compute.manager [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 978.481016] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4125848-eaff-46a5-87a2-b69f5bbc2cf1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.577440] env[62385]: DEBUG oslo_vmware.api [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206245, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.748387] env[62385]: DEBUG nova.compute.manager [req-9b2ba4a0-37ce-429f-934f-026405a548b8 req-36053cbd-7cbb-4578-ab25-c641fb8b3ba1 service nova] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Received event network-vif-deleted-31a45e90-bf05-4452-8a2c-490aa557ac1b {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 978.748450] env[62385]: INFO nova.compute.manager [req-9b2ba4a0-37ce-429f-934f-026405a548b8 req-36053cbd-7cbb-4578-ab25-c641fb8b3ba1 service nova] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Neutron deleted interface 31a45e90-bf05-4452-8a2c-490aa557ac1b; detaching it from the instance and deleting it from the info cache [ 978.748626] env[62385]: DEBUG nova.network.neutron [req-9b2ba4a0-37ce-429f-934f-026405a548b8 req-36053cbd-7cbb-4578-ab25-c641fb8b3ba1 service nova] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.832112] env[62385]: DEBUG nova.scheduler.client.report [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 979.003299] env[62385]: INFO nova.compute.manager [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Took 24.71 seconds to build instance. [ 979.077020] env[62385]: DEBUG oslo_vmware.api [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206245, 'name': PowerOnVM_Task, 'duration_secs': 0.521968} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.077457] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 979.077556] env[62385]: INFO nova.compute.manager [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Took 7.37 seconds to spawn the instance on the hypervisor. [ 979.077693] env[62385]: DEBUG nova.compute.manager [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 979.078468] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2046df43-9490-48f6-b1f3-1636fe333ff3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.217604] env[62385]: DEBUG nova.network.neutron [-] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.253168] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b4add5d4-de40-41a8-9ff9-6607dbcc9b32 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.270626] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a355388-2edc-47a4-9269-1e399ce0c485 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.300691] env[62385]: DEBUG nova.compute.manager [req-9b2ba4a0-37ce-429f-934f-026405a548b8 req-36053cbd-7cbb-4578-ab25-c641fb8b3ba1 service nova] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Detach interface failed, port_id=31a45e90-bf05-4452-8a2c-490aa557ac1b, reason: Instance fd0b59c2-e38d-4ecc-b32f-72f07d555ffd could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 979.305833] env[62385]: DEBUG oslo_concurrency.lockutils [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "76672404-2bcf-438a-9e21-92ea9dc86461" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.306213] env[62385]: DEBUG oslo_concurrency.lockutils [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "76672404-2bcf-438a-9e21-92ea9dc86461" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.337717] env[62385]: DEBUG oslo_concurrency.lockutils [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.343s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.337968] env[62385]: INFO nova.compute.manager [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Migrating [ 979.345386] env[62385]: DEBUG oslo_concurrency.lockutils [None req-74f8a615-f91b-4cab-a95b-682654db78d4 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.672s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.345627] env[62385]: DEBUG nova.objects.instance [None req-74f8a615-f91b-4cab-a95b-682654db78d4 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lazy-loading 'resources' on Instance uuid f937ef7b-ab39-4d8a-9577-7faabc652aa5 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 979.505574] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2de57272-716f-4a1b-a74a-97cc3e2f3320 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "9741e9c8-60d9-41bf-8185-ef3cfcc1c844" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.218s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.599311] env[62385]: INFO nova.compute.manager [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Took 24.83 seconds to build instance. [ 979.720811] env[62385]: INFO nova.compute.manager [-] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Took 1.59 seconds to deallocate network for instance. [ 979.808755] env[62385]: DEBUG nova.compute.manager [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 979.857574] env[62385]: DEBUG oslo_concurrency.lockutils [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "refresh_cache-df020d6b-3fab-4599-a342-47c7833b4240" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.857574] env[62385]: DEBUG oslo_concurrency.lockutils [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquired lock "refresh_cache-df020d6b-3fab-4599-a342-47c7833b4240" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.857574] env[62385]: DEBUG nova.network.neutron [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 980.053023] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-397a696b-40dd-4953-bcf3-8041fe772233 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.061571] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-433a7869-276f-4c90-ae79-3e53340a3375 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.093110] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27d16bc9-e56e-4062-b6fc-f947775b7357 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.101469] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c61a1754-983f-437a-8ff3-225da759fb79 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "a1c71d8b-0953-4cc2-961d-0c2b2fc11110" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.343s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.106019] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86158f6f-af77-4de6-bbfd-ff6963aedb78 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.119452] env[62385]: DEBUG nova.compute.provider_tree [None req-74f8a615-f91b-4cab-a95b-682654db78d4 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 980.229454] env[62385]: DEBUG oslo_concurrency.lockutils [None req-73f55945-68c9-4e01-b622-3cd3be847fe4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.336466] env[62385]: DEBUG oslo_concurrency.lockutils [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.358248] env[62385]: DEBUG nova.compute.manager [req-4b5e3ea3-8bd0-495f-8030-edbc09a55b82 req-278076c2-9ca0-4ea0-a0ea-0fe1efce0b21 service nova] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Received event network-changed-20926145-a8ac-4f6f-afb1-a1c449d1d76d {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 980.358494] env[62385]: DEBUG nova.compute.manager [req-4b5e3ea3-8bd0-495f-8030-edbc09a55b82 req-278076c2-9ca0-4ea0-a0ea-0fe1efce0b21 service nova] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Refreshing instance network info cache due to event network-changed-20926145-a8ac-4f6f-afb1-a1c449d1d76d. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 980.358620] env[62385]: DEBUG oslo_concurrency.lockutils [req-4b5e3ea3-8bd0-495f-8030-edbc09a55b82 req-278076c2-9ca0-4ea0-a0ea-0fe1efce0b21 service nova] Acquiring lock "refresh_cache-a1c71d8b-0953-4cc2-961d-0c2b2fc11110" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.358747] env[62385]: DEBUG oslo_concurrency.lockutils [req-4b5e3ea3-8bd0-495f-8030-edbc09a55b82 req-278076c2-9ca0-4ea0-a0ea-0fe1efce0b21 service nova] Acquired lock "refresh_cache-a1c71d8b-0953-4cc2-961d-0c2b2fc11110" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.358915] env[62385]: DEBUG nova.network.neutron [req-4b5e3ea3-8bd0-495f-8030-edbc09a55b82 req-278076c2-9ca0-4ea0-a0ea-0fe1efce0b21 service nova] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Refreshing network info cache for port 20926145-a8ac-4f6f-afb1-a1c449d1d76d {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 980.622287] env[62385]: DEBUG nova.scheduler.client.report [None req-74f8a615-f91b-4cab-a95b-682654db78d4 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 980.721937] env[62385]: DEBUG nova.compute.manager [None req-088beef8-dd2f-4e4a-b0b0-98f5d7544997 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 980.723032] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcd3a9ac-2ca0-4cb7-a8ed-c2e8f682151f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.800432] env[62385]: DEBUG nova.network.neutron [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Updating instance_info_cache with network_info: [{"id": "a9513285-4023-4f3e-b90d-3eb507955f13", "address": "fa:16:3e:30:02:c0", "network": {"id": "30c4e78b-6a64-47dc-aaaa-a210160c9f23", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1809587656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "105150c170d74e268e82ab894e9bf0d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10ff2092-e8eb-4768-ad4a-65a80560b447", "external-id": "nsx-vlan-transportzone-845", "segmentation_id": 845, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa9513285-40", "ovs_interfaceid": "a9513285-4023-4f3e-b90d-3eb507955f13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.128269] env[62385]: DEBUG oslo_concurrency.lockutils [None req-74f8a615-f91b-4cab-a95b-682654db78d4 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.782s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.130151] env[62385]: DEBUG oslo_concurrency.lockutils [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 5.575s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.130365] env[62385]: DEBUG nova.objects.instance [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62385) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 981.156799] env[62385]: INFO nova.scheduler.client.report [None req-74f8a615-f91b-4cab-a95b-682654db78d4 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Deleted allocations for instance f937ef7b-ab39-4d8a-9577-7faabc652aa5 [ 981.235164] env[62385]: INFO nova.compute.manager [None req-088beef8-dd2f-4e4a-b0b0-98f5d7544997 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] instance snapshotting [ 981.238282] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb7eb4f3-4e49-423a-83d1-8af52cc09ec5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.260012] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f77dc3df-ed92-4f89-93ee-dabaa401b66f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.309434] env[62385]: DEBUG oslo_concurrency.lockutils [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Releasing lock "refresh_cache-df020d6b-3fab-4599-a342-47c7833b4240" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.414338] env[62385]: DEBUG nova.network.neutron [req-4b5e3ea3-8bd0-495f-8030-edbc09a55b82 req-278076c2-9ca0-4ea0-a0ea-0fe1efce0b21 service nova] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Updated VIF entry in instance network info cache for port 20926145-a8ac-4f6f-afb1-a1c449d1d76d. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 981.414338] env[62385]: DEBUG nova.network.neutron [req-4b5e3ea3-8bd0-495f-8030-edbc09a55b82 req-278076c2-9ca0-4ea0-a0ea-0fe1efce0b21 service nova] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Updating instance_info_cache with network_info: [{"id": "20926145-a8ac-4f6f-afb1-a1c449d1d76d", "address": "fa:16:3e:98:24:99", "network": {"id": "aca904e1-fbd6-4def-bbec-bd70c0fc8af9", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-249813742-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.176", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e99f1e4cfd84bf083c8261b2ffcfa2f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f7abe5d-b8fe-4983-bd50-e7469f1fe7f3", "external-id": "nsx-vlan-transportzone-263", "segmentation_id": 263, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap20926145-a8", "ovs_interfaceid": "20926145-a8ac-4f6f-afb1-a1c449d1d76d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.669789] env[62385]: DEBUG oslo_concurrency.lockutils [None req-74f8a615-f91b-4cab-a95b-682654db78d4 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "f937ef7b-ab39-4d8a-9577-7faabc652aa5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.194s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.773568] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-088beef8-dd2f-4e4a-b0b0-98f5d7544997 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Creating Snapshot of the VM instance {{(pid=62385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 981.773568] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-cc242265-d53d-4474-aade-4ee5dda6c57d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.782480] env[62385]: DEBUG oslo_vmware.api [None req-088beef8-dd2f-4e4a-b0b0-98f5d7544997 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 981.782480] env[62385]: value = "task-1206246" [ 981.782480] env[62385]: _type = "Task" [ 981.782480] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.791518] env[62385]: DEBUG oslo_vmware.api [None req-088beef8-dd2f-4e4a-b0b0-98f5d7544997 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206246, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.917128] env[62385]: DEBUG oslo_concurrency.lockutils [req-4b5e3ea3-8bd0-495f-8030-edbc09a55b82 req-278076c2-9ca0-4ea0-a0ea-0fe1efce0b21 service nova] Releasing lock "refresh_cache-a1c71d8b-0953-4cc2-961d-0c2b2fc11110" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.138880] env[62385]: DEBUG oslo_concurrency.lockutils [None req-06a55ea8-0ffb-4af0-b861-ecc47a66d867 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.009s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.140057] env[62385]: DEBUG oslo_concurrency.lockutils [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 5.130s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.293802] env[62385]: DEBUG oslo_vmware.api [None req-088beef8-dd2f-4e4a-b0b0-98f5d7544997 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206246, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.650634] env[62385]: INFO nova.compute.claims [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 982.794475] env[62385]: DEBUG oslo_vmware.api [None req-088beef8-dd2f-4e4a-b0b0-98f5d7544997 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206246, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.826934] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ea5bd42-d427-4f69-9dad-6fcbea522130 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.851266] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Updating instance 'df020d6b-3fab-4599-a342-47c7833b4240' progress to 0 {{(pid=62385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 983.119297] env[62385]: DEBUG oslo_concurrency.lockutils [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Acquiring lock "784589bb-a668-4674-83b8-6219def15067" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.119297] env[62385]: DEBUG oslo_concurrency.lockutils [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Lock "784589bb-a668-4674-83b8-6219def15067" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.159021] env[62385]: INFO nova.compute.resource_tracker [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Updating resource usage from migration 103f026b-ef52-4238-8a48-c1a1b0aed1b6 [ 983.297080] env[62385]: DEBUG oslo_vmware.api [None req-088beef8-dd2f-4e4a-b0b0-98f5d7544997 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206246, 'name': CreateSnapshot_Task, 'duration_secs': 1.127755} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.297080] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-088beef8-dd2f-4e4a-b0b0-98f5d7544997 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Created Snapshot of the VM instance {{(pid=62385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 983.297080] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de701cc5-271c-4205-8abf-356983644f88 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.363153] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 983.365830] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e9675d04-17be-4b9b-a06d-c3a6a5baa63e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.374034] env[62385]: DEBUG oslo_vmware.api [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 983.374034] env[62385]: value = "task-1206247" [ 983.374034] env[62385]: _type = "Task" [ 983.374034] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.385786] env[62385]: DEBUG oslo_vmware.api [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206247, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.411389] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63250403-69d5-4e85-a6ab-0d9e594dd455 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.424084] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6dc5c09-13a0-430b-b46b-a4951dd82994 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.472352] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24af604f-e858-4817-bcfa-be5b4caea124 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.483381] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eea1332-6105-4ff6-9a3a-95ecce3140d0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.505158] env[62385]: DEBUG nova.compute.provider_tree [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 983.622360] env[62385]: DEBUG nova.compute.manager [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] [instance: 784589bb-a668-4674-83b8-6219def15067] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 983.817447] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-088beef8-dd2f-4e4a-b0b0-98f5d7544997 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Creating linked-clone VM from snapshot {{(pid=62385) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 983.817883] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-d9cc68c2-de86-4f98-b6da-c689dfb3e97c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.828015] env[62385]: DEBUG oslo_vmware.api [None req-088beef8-dd2f-4e4a-b0b0-98f5d7544997 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 983.828015] env[62385]: value = "task-1206248" [ 983.828015] env[62385]: _type = "Task" [ 983.828015] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.836569] env[62385]: DEBUG oslo_vmware.api [None req-088beef8-dd2f-4e4a-b0b0-98f5d7544997 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206248, 'name': CloneVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.882716] env[62385]: DEBUG oslo_vmware.api [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206247, 'name': PowerOffVM_Task, 'duration_secs': 0.224956} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.883024] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 983.883231] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Updating instance 'df020d6b-3fab-4599-a342-47c7833b4240' progress to 17 {{(pid=62385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 984.009968] env[62385]: DEBUG nova.scheduler.client.report [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 984.075642] env[62385]: DEBUG oslo_concurrency.lockutils [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "57637d6e-39e2-405c-bbcf-64bf0ffbebd1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.075843] env[62385]: DEBUG oslo_concurrency.lockutils [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "57637d6e-39e2-405c-bbcf-64bf0ffbebd1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.152050] env[62385]: DEBUG oslo_concurrency.lockutils [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.338898] env[62385]: DEBUG oslo_vmware.api [None req-088beef8-dd2f-4e4a-b0b0-98f5d7544997 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206248, 'name': CloneVM_Task} progress is 94%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.389904] env[62385]: DEBUG nova.virt.hardware [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:16:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 984.390221] env[62385]: DEBUG nova.virt.hardware [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 984.390394] env[62385]: DEBUG nova.virt.hardware [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 984.390581] env[62385]: DEBUG nova.virt.hardware [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 984.390735] env[62385]: DEBUG nova.virt.hardware [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 984.390891] env[62385]: DEBUG nova.virt.hardware [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 984.391406] env[62385]: DEBUG nova.virt.hardware [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 984.391608] env[62385]: DEBUG nova.virt.hardware [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 984.391791] env[62385]: DEBUG nova.virt.hardware [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 984.392239] env[62385]: DEBUG nova.virt.hardware [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 984.392668] env[62385]: DEBUG nova.virt.hardware [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 984.398522] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f13b0245-49af-4abd-ac6e-1f869f6fa742 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.417545] env[62385]: DEBUG oslo_vmware.api [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 984.417545] env[62385]: value = "task-1206249" [ 984.417545] env[62385]: _type = "Task" [ 984.417545] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.427729] env[62385]: DEBUG oslo_vmware.api [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206249, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.516036] env[62385]: DEBUG oslo_concurrency.lockutils [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.376s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.516280] env[62385]: INFO nova.compute.manager [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Migrating [ 984.523553] env[62385]: DEBUG oslo_concurrency.lockutils [None req-73f55945-68c9-4e01-b622-3cd3be847fe4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.294s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.523825] env[62385]: DEBUG nova.objects.instance [None req-73f55945-68c9-4e01-b622-3cd3be847fe4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lazy-loading 'resources' on Instance uuid fd0b59c2-e38d-4ecc-b32f-72f07d555ffd {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 984.578667] env[62385]: DEBUG nova.compute.manager [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 984.839801] env[62385]: DEBUG oslo_vmware.api [None req-088beef8-dd2f-4e4a-b0b0-98f5d7544997 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206248, 'name': CloneVM_Task} progress is 94%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.928717] env[62385]: DEBUG oslo_vmware.api [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206249, 'name': ReconfigVM_Task, 'duration_secs': 0.305995} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.931820] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Updating instance 'df020d6b-3fab-4599-a342-47c7833b4240' progress to 33 {{(pid=62385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 985.036049] env[62385]: DEBUG oslo_concurrency.lockutils [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "refresh_cache-b1e0baf3-643c-49c2-8a80-8de07d8527e7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.036249] env[62385]: DEBUG oslo_concurrency.lockutils [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquired lock "refresh_cache-b1e0baf3-643c-49c2-8a80-8de07d8527e7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.036420] env[62385]: DEBUG nova.network.neutron [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 985.104052] env[62385]: DEBUG oslo_concurrency.lockutils [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.251186] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c494d285-3a3e-4e95-a588-9fb8ccaa25b3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.259174] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0bf7e90-d7d3-4574-8893-cdf033a7e040 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.292865] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6860f61c-717e-4d56-bbb7-ab9458c3d391 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.299684] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquiring lock "8e1d416c-2624-468c-94d7-6265e9f4178f" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.299972] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lock "8e1d416c-2624-468c-94d7-6265e9f4178f" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.300171] env[62385]: INFO nova.compute.manager [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Shelving [ 985.302551] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1fe5a73-e80b-46a4-a9ad-049dcd64a1c2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.317267] env[62385]: DEBUG nova.compute.provider_tree [None req-73f55945-68c9-4e01-b622-3cd3be847fe4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 985.338884] env[62385]: DEBUG oslo_vmware.api [None req-088beef8-dd2f-4e4a-b0b0-98f5d7544997 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206248, 'name': CloneVM_Task} progress is 95%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.435662] env[62385]: DEBUG nova.virt.hardware [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 985.435918] env[62385]: DEBUG nova.virt.hardware [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 985.436104] env[62385]: DEBUG nova.virt.hardware [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 985.436289] env[62385]: DEBUG nova.virt.hardware [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 985.436441] env[62385]: DEBUG nova.virt.hardware [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 985.436593] env[62385]: DEBUG nova.virt.hardware [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 985.436800] env[62385]: DEBUG nova.virt.hardware [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 985.437029] env[62385]: DEBUG nova.virt.hardware [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 985.437149] env[62385]: DEBUG nova.virt.hardware [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 985.437323] env[62385]: DEBUG nova.virt.hardware [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 985.437498] env[62385]: DEBUG nova.virt.hardware [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 985.442877] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Reconfiguring VM instance instance-00000060 to detach disk 2000 {{(pid=62385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 985.443509] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b7142cc0-2e9c-4d3e-809a-d069ec13106f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.462105] env[62385]: DEBUG oslo_vmware.api [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 985.462105] env[62385]: value = "task-1206250" [ 985.462105] env[62385]: _type = "Task" [ 985.462105] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.470290] env[62385]: DEBUG oslo_vmware.api [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206250, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.738861] env[62385]: DEBUG nova.network.neutron [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Updating instance_info_cache with network_info: [{"id": "eef524f4-fc00-47f1-a485-dcf568c6d0e4", "address": "fa:16:3e:e6:2c:55", "network": {"id": "ff21dd88-fa02-4bd6-8009-980ec6e72758", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1065789169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aea48968d4464ac0ab01c998c059107d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3ff3baee-99ce-4b51-ae98-efc6163aaab3", "external-id": "nsx-vlan-transportzone-574", "segmentation_id": 574, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeef524f4-fc", "ovs_interfaceid": "eef524f4-fc00-47f1-a485-dcf568c6d0e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.812685] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 985.812956] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0776a13a-874d-471f-ba9b-6871971b3f36 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.819989] env[62385]: DEBUG nova.scheduler.client.report [None req-73f55945-68c9-4e01-b622-3cd3be847fe4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 985.823744] env[62385]: DEBUG oslo_vmware.api [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 985.823744] env[62385]: value = "task-1206251" [ 985.823744] env[62385]: _type = "Task" [ 985.823744] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.832817] env[62385]: DEBUG oslo_vmware.api [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206251, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.841130] env[62385]: DEBUG oslo_vmware.api [None req-088beef8-dd2f-4e4a-b0b0-98f5d7544997 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206248, 'name': CloneVM_Task, 'duration_secs': 1.67208} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.841390] env[62385]: INFO nova.virt.vmwareapi.vmops [None req-088beef8-dd2f-4e4a-b0b0-98f5d7544997 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Created linked-clone VM from snapshot [ 985.842132] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-665b5d51-dc29-4170-b9ba-3850a69bbb9d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.850166] env[62385]: DEBUG nova.virt.vmwareapi.images [None req-088beef8-dd2f-4e4a-b0b0-98f5d7544997 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Uploading image 4ea01292-abb8-4034-b95c-8c42041f7d73 {{(pid=62385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 985.862843] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-088beef8-dd2f-4e4a-b0b0-98f5d7544997 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Destroying the VM {{(pid=62385) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 985.863368] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-e3ba737c-b5af-41fa-a74a-1a0b07c0f65f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.869502] env[62385]: DEBUG oslo_vmware.api [None req-088beef8-dd2f-4e4a-b0b0-98f5d7544997 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 985.869502] env[62385]: value = "task-1206252" [ 985.869502] env[62385]: _type = "Task" [ 985.869502] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.877641] env[62385]: DEBUG oslo_vmware.api [None req-088beef8-dd2f-4e4a-b0b0-98f5d7544997 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206252, 'name': Destroy_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.973016] env[62385]: DEBUG oslo_vmware.api [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206250, 'name': ReconfigVM_Task, 'duration_secs': 0.436714} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.973319] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Reconfigured VM instance instance-00000060 to detach disk 2000 {{(pid=62385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 985.974116] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e01b0b93-1698-4cc9-ba4a-1389d6fe218a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.998476] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] df020d6b-3fab-4599-a342-47c7833b4240/df020d6b-3fab-4599-a342-47c7833b4240.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 985.998983] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-57b2df62-3a5e-482c-85c1-c6a92e4af31a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.018160] env[62385]: DEBUG oslo_vmware.api [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 986.018160] env[62385]: value = "task-1206253" [ 986.018160] env[62385]: _type = "Task" [ 986.018160] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.027239] env[62385]: DEBUG oslo_vmware.api [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206253, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.241741] env[62385]: DEBUG oslo_concurrency.lockutils [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Releasing lock "refresh_cache-b1e0baf3-643c-49c2-8a80-8de07d8527e7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.325352] env[62385]: DEBUG oslo_concurrency.lockutils [None req-73f55945-68c9-4e01-b622-3cd3be847fe4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.802s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.328374] env[62385]: DEBUG oslo_concurrency.lockutils [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.992s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.329796] env[62385]: INFO nova.compute.claims [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 986.343666] env[62385]: DEBUG oslo_vmware.api [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206251, 'name': PowerOffVM_Task, 'duration_secs': 0.172983} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.343992] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 986.345183] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2cd834e-42e3-4330-b9b8-8cfb2e417b8f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.370672] env[62385]: INFO nova.scheduler.client.report [None req-73f55945-68c9-4e01-b622-3cd3be847fe4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Deleted allocations for instance fd0b59c2-e38d-4ecc-b32f-72f07d555ffd [ 986.378296] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5ba6560-0535-48e9-b27d-047b06b4c847 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.390261] env[62385]: DEBUG oslo_vmware.api [None req-088beef8-dd2f-4e4a-b0b0-98f5d7544997 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206252, 'name': Destroy_Task, 'duration_secs': 0.302888} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.392055] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-088beef8-dd2f-4e4a-b0b0-98f5d7544997 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Destroyed the VM [ 986.393037] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-088beef8-dd2f-4e4a-b0b0-98f5d7544997 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Deleting Snapshot of the VM instance {{(pid=62385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 986.395618] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-c81ad4bc-96ad-46ca-8ac3-9ee6cd093f07 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.403925] env[62385]: DEBUG oslo_vmware.api [None req-088beef8-dd2f-4e4a-b0b0-98f5d7544997 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 986.403925] env[62385]: value = "task-1206254" [ 986.403925] env[62385]: _type = "Task" [ 986.403925] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.412740] env[62385]: DEBUG oslo_vmware.api [None req-088beef8-dd2f-4e4a-b0b0-98f5d7544997 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206254, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.528802] env[62385]: DEBUG oslo_vmware.api [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206253, 'name': ReconfigVM_Task, 'duration_secs': 0.270013} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.530690] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Reconfigured VM instance instance-00000060 to attach disk [datastore2] df020d6b-3fab-4599-a342-47c7833b4240/df020d6b-3fab-4599-a342-47c7833b4240.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 986.530690] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Updating instance 'df020d6b-3fab-4599-a342-47c7833b4240' progress to 50 {{(pid=62385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 986.882302] env[62385]: DEBUG oslo_concurrency.lockutils [None req-73f55945-68c9-4e01-b622-3cd3be847fe4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "fd0b59c2-e38d-4ecc-b32f-72f07d555ffd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.421s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.897661] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Creating Snapshot of the VM instance {{(pid=62385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 986.898532] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-ba109550-72e4-4dd5-8848-e8b6b5457be9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.909660] env[62385]: DEBUG oslo_vmware.api [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 986.909660] env[62385]: value = "task-1206255" [ 986.909660] env[62385]: _type = "Task" [ 986.909660] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.918094] env[62385]: DEBUG oslo_vmware.api [None req-088beef8-dd2f-4e4a-b0b0-98f5d7544997 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206254, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.922753] env[62385]: DEBUG oslo_vmware.api [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206255, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.036639] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bad2b265-0982-4274-998e-ca470c0bf732 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.056930] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57e3fde7-76b8-4522-965f-411ccc9fa3db {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.076218] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Updating instance 'df020d6b-3fab-4599-a342-47c7833b4240' progress to 67 {{(pid=62385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 987.420801] env[62385]: DEBUG oslo_vmware.api [None req-088beef8-dd2f-4e4a-b0b0-98f5d7544997 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206254, 'name': RemoveSnapshot_Task, 'duration_secs': 1.014996} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.423952] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-088beef8-dd2f-4e4a-b0b0-98f5d7544997 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Deleted Snapshot of the VM instance {{(pid=62385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 987.426612] env[62385]: DEBUG oslo_vmware.api [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206255, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.566014] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c8d8b9a-ccdc-4602-be39-c92860e2889b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.575971] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f46afb9-1015-452d-b8a0-3a2c3db04c47 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.618807] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ca7b723-ce24-4d7e-82a6-efe7b07c0e5f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.626408] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e284ae81-45b3-408a-bed6-9684d4768fe1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.639545] env[62385]: DEBUG nova.compute.provider_tree [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 987.665848] env[62385]: DEBUG nova.network.neutron [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Port a9513285-4023-4f3e-b90d-3eb507955f13 binding to destination host cpu-1 is already ACTIVE {{(pid=62385) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 987.759565] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05c74c17-88c8-48a3-aab0-78fb26932b27 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.778308] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Updating instance 'b1e0baf3-643c-49c2-8a80-8de07d8527e7' progress to 0 {{(pid=62385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 987.922132] env[62385]: DEBUG oslo_vmware.api [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206255, 'name': CreateSnapshot_Task, 'duration_secs': 0.806547} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.922449] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Created Snapshot of the VM instance {{(pid=62385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 987.923666] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e21b0581-ffd1-4f16-9ab4-08924a4b4d71 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.928281] env[62385]: WARNING nova.compute.manager [None req-088beef8-dd2f-4e4a-b0b0-98f5d7544997 tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Image not found during snapshot: nova.exception.ImageNotFound: Image 4ea01292-abb8-4034-b95c-8c42041f7d73 could not be found. [ 988.142550] env[62385]: DEBUG nova.scheduler.client.report [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 988.287010] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 988.287354] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f4257340-3df4-4703-b3ac-6f43a2b8b8cf {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.295259] env[62385]: DEBUG oslo_vmware.api [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 988.295259] env[62385]: value = "task-1206256" [ 988.295259] env[62385]: _type = "Task" [ 988.295259] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.306236] env[62385]: DEBUG oslo_vmware.api [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206256, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.447725] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Creating linked-clone VM from snapshot {{(pid=62385) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 988.448661] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-0d5f282a-83c5-4a00-a353-404c22de7aa4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.458682] env[62385]: DEBUG oslo_vmware.api [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 988.458682] env[62385]: value = "task-1206257" [ 988.458682] env[62385]: _type = "Task" [ 988.458682] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.467866] env[62385]: DEBUG oslo_vmware.api [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206257, 'name': CloneVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.649950] env[62385]: DEBUG oslo_concurrency.lockutils [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.322s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.650722] env[62385]: DEBUG nova.compute.manager [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 988.654600] env[62385]: DEBUG oslo_concurrency.lockutils [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.503s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.656751] env[62385]: INFO nova.compute.claims [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] [instance: 784589bb-a668-4674-83b8-6219def15067] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 988.696031] env[62385]: DEBUG oslo_concurrency.lockutils [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "df020d6b-3fab-4599-a342-47c7833b4240-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.696031] env[62385]: DEBUG oslo_concurrency.lockutils [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "df020d6b-3fab-4599-a342-47c7833b4240-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.696031] env[62385]: DEBUG oslo_concurrency.lockutils [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "df020d6b-3fab-4599-a342-47c7833b4240-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.806171] env[62385]: DEBUG oslo_vmware.api [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206256, 'name': PowerOffVM_Task, 'duration_secs': 0.187223} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.809898] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 988.809898] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Updating instance 'b1e0baf3-643c-49c2-8a80-8de07d8527e7' progress to 17 {{(pid=62385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 988.878884] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a3a6fc1a-c0ca-4fe5-9a14-65b9435d0e5f tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquiring lock "9741e9c8-60d9-41bf-8185-ef3cfcc1c844" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.879738] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a3a6fc1a-c0ca-4fe5-9a14-65b9435d0e5f tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "9741e9c8-60d9-41bf-8185-ef3cfcc1c844" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.879738] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a3a6fc1a-c0ca-4fe5-9a14-65b9435d0e5f tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquiring lock "9741e9c8-60d9-41bf-8185-ef3cfcc1c844-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.879738] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a3a6fc1a-c0ca-4fe5-9a14-65b9435d0e5f tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "9741e9c8-60d9-41bf-8185-ef3cfcc1c844-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.879738] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a3a6fc1a-c0ca-4fe5-9a14-65b9435d0e5f tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "9741e9c8-60d9-41bf-8185-ef3cfcc1c844-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.882047] env[62385]: INFO nova.compute.manager [None req-a3a6fc1a-c0ca-4fe5-9a14-65b9435d0e5f tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Terminating instance [ 988.884203] env[62385]: DEBUG nova.compute.manager [None req-a3a6fc1a-c0ca-4fe5-9a14-65b9435d0e5f tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 988.884203] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a3a6fc1a-c0ca-4fe5-9a14-65b9435d0e5f tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 988.885096] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0afa4a6-1ca5-496f-8197-26e60d6921bb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.893614] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3a6fc1a-c0ca-4fe5-9a14-65b9435d0e5f tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 988.893614] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2474b58c-bf8a-4b9d-a161-4f92dcb3ec19 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.899838] env[62385]: DEBUG oslo_vmware.api [None req-a3a6fc1a-c0ca-4fe5-9a14-65b9435d0e5f tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 988.899838] env[62385]: value = "task-1206258" [ 988.899838] env[62385]: _type = "Task" [ 988.899838] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.908329] env[62385]: DEBUG oslo_vmware.api [None req-a3a6fc1a-c0ca-4fe5-9a14-65b9435d0e5f tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206258, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.968538] env[62385]: DEBUG oslo_vmware.api [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206257, 'name': CloneVM_Task} progress is 94%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.161585] env[62385]: DEBUG nova.compute.utils [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 989.165366] env[62385]: DEBUG nova.compute.manager [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 989.165558] env[62385]: DEBUG nova.network.neutron [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 989.220813] env[62385]: DEBUG nova.policy [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5e83312d226945c99d05fcc7f74c0978', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bf8e3c942d3445919cfbe988cca84e90', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 989.224984] env[62385]: DEBUG nova.compute.manager [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Stashing vm_state: active {{(pid=62385) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 989.313184] env[62385]: DEBUG nova.virt.hardware [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:16:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 989.313476] env[62385]: DEBUG nova.virt.hardware [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 989.313651] env[62385]: DEBUG nova.virt.hardware [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 989.313841] env[62385]: DEBUG nova.virt.hardware [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 989.314407] env[62385]: DEBUG nova.virt.hardware [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 989.314407] env[62385]: DEBUG nova.virt.hardware [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 989.314407] env[62385]: DEBUG nova.virt.hardware [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 989.314573] env[62385]: DEBUG nova.virt.hardware [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 989.314745] env[62385]: DEBUG nova.virt.hardware [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 989.314956] env[62385]: DEBUG nova.virt.hardware [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 989.315189] env[62385]: DEBUG nova.virt.hardware [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 989.320978] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-107e8dc5-5b06-42aa-8804-c0762e9d0b93 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.336957] env[62385]: DEBUG oslo_vmware.api [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 989.336957] env[62385]: value = "task-1206259" [ 989.336957] env[62385]: _type = "Task" [ 989.336957] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.345443] env[62385]: DEBUG oslo_vmware.api [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206259, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.410432] env[62385]: DEBUG oslo_vmware.api [None req-a3a6fc1a-c0ca-4fe5-9a14-65b9435d0e5f tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206258, 'name': PowerOffVM_Task, 'duration_secs': 0.411085} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.410756] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3a6fc1a-c0ca-4fe5-9a14-65b9435d0e5f tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 989.411205] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a3a6fc1a-c0ca-4fe5-9a14-65b9435d0e5f tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 989.411532] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a87afb1d-382e-43a3-adbb-82e840bcc1ca {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.470048] env[62385]: DEBUG oslo_vmware.api [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206257, 'name': CloneVM_Task} progress is 94%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.478362] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a3a6fc1a-c0ca-4fe5-9a14-65b9435d0e5f tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 989.478599] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a3a6fc1a-c0ca-4fe5-9a14-65b9435d0e5f tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Deleting contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 989.478792] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3a6fc1a-c0ca-4fe5-9a14-65b9435d0e5f tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Deleting the datastore file [datastore2] 9741e9c8-60d9-41bf-8185-ef3cfcc1c844 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 989.479078] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-52195801-a50c-40bd-990c-e65d8cd4b791 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.484986] env[62385]: DEBUG nova.network.neutron [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Successfully created port: 117fbbdf-9bdd-4d68-ac9f-d2fa6279aebd {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 989.488524] env[62385]: DEBUG oslo_vmware.api [None req-a3a6fc1a-c0ca-4fe5-9a14-65b9435d0e5f tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for the task: (returnval){ [ 989.488524] env[62385]: value = "task-1206261" [ 989.488524] env[62385]: _type = "Task" [ 989.488524] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.497012] env[62385]: DEBUG oslo_vmware.api [None req-a3a6fc1a-c0ca-4fe5-9a14-65b9435d0e5f tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206261, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.668775] env[62385]: DEBUG nova.compute.manager [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 989.748815] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.787114] env[62385]: DEBUG oslo_concurrency.lockutils [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "refresh_cache-df020d6b-3fab-4599-a342-47c7833b4240" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.787114] env[62385]: DEBUG oslo_concurrency.lockutils [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquired lock "refresh_cache-df020d6b-3fab-4599-a342-47c7833b4240" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.787114] env[62385]: DEBUG nova.network.neutron [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 989.851216] env[62385]: DEBUG oslo_vmware.api [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206259, 'name': ReconfigVM_Task, 'duration_secs': 0.2462} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.851216] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Updating instance 'b1e0baf3-643c-49c2-8a80-8de07d8527e7' progress to 33 {{(pid=62385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 989.930020] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c4e160e-23d9-4a98-9839-9698f024ab7e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.936760] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7e4c9b5-46fa-4787-947b-8874f6205b0b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.970105] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7ac7de2-eda4-4efc-a5ad-ff6cfcd4a142 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.981854] env[62385]: DEBUG oslo_vmware.api [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206257, 'name': CloneVM_Task, 'duration_secs': 1.417134} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.984620] env[62385]: INFO nova.virt.vmwareapi.vmops [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Created linked-clone VM from snapshot [ 989.985489] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec79b3f9-dc23-4162-a0fa-68eadf59c401 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.989399] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25d2e757-6375-4d6f-8c80-6caea1a11d17 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.002058] env[62385]: DEBUG nova.virt.vmwareapi.images [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Uploading image a561779b-d9f7-4373-8a4d-586fbc25c6d7 {{(pid=62385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 990.015658] env[62385]: DEBUG nova.compute.provider_tree [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 990.021884] env[62385]: DEBUG oslo_vmware.api [None req-a3a6fc1a-c0ca-4fe5-9a14-65b9435d0e5f tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Task: {'id': task-1206261, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.142965} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.025026] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3a6fc1a-c0ca-4fe5-9a14-65b9435d0e5f tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 990.025026] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a3a6fc1a-c0ca-4fe5-9a14-65b9435d0e5f tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Deleted contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 990.025026] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a3a6fc1a-c0ca-4fe5-9a14-65b9435d0e5f tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 990.025026] env[62385]: INFO nova.compute.manager [None req-a3a6fc1a-c0ca-4fe5-9a14-65b9435d0e5f tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Took 1.14 seconds to destroy the instance on the hypervisor. [ 990.025026] env[62385]: DEBUG oslo.service.loopingcall [None req-a3a6fc1a-c0ca-4fe5-9a14-65b9435d0e5f tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 990.025026] env[62385]: DEBUG nova.compute.manager [-] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 990.025026] env[62385]: DEBUG nova.network.neutron [-] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 990.040861] env[62385]: DEBUG oslo_vmware.rw_handles [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 990.040861] env[62385]: value = "vm-261277" [ 990.040861] env[62385]: _type = "VirtualMachine" [ 990.040861] env[62385]: }. {{(pid=62385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 990.041141] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-2157cdf1-5560-40a2-9f86-dd1f5438db3d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.048715] env[62385]: DEBUG oslo_vmware.rw_handles [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lease: (returnval){ [ 990.048715] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e4ff39-fc59-5b58-ec9e-cb3a9fb0730c" [ 990.048715] env[62385]: _type = "HttpNfcLease" [ 990.048715] env[62385]: } obtained for exporting VM: (result){ [ 990.048715] env[62385]: value = "vm-261277" [ 990.048715] env[62385]: _type = "VirtualMachine" [ 990.048715] env[62385]: }. {{(pid=62385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 990.048939] env[62385]: DEBUG oslo_vmware.api [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the lease: (returnval){ [ 990.048939] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e4ff39-fc59-5b58-ec9e-cb3a9fb0730c" [ 990.048939] env[62385]: _type = "HttpNfcLease" [ 990.048939] env[62385]: } to be ready. {{(pid=62385) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 990.055739] env[62385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 990.055739] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e4ff39-fc59-5b58-ec9e-cb3a9fb0730c" [ 990.055739] env[62385]: _type = "HttpNfcLease" [ 990.055739] env[62385]: } is initializing. {{(pid=62385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 990.356835] env[62385]: DEBUG nova.virt.hardware [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 990.358920] env[62385]: DEBUG nova.virt.hardware [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 990.358920] env[62385]: DEBUG nova.virt.hardware [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 990.358920] env[62385]: DEBUG nova.virt.hardware [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 990.358920] env[62385]: DEBUG nova.virt.hardware [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 990.358920] env[62385]: DEBUG nova.virt.hardware [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 990.358920] env[62385]: DEBUG nova.virt.hardware [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 990.358920] env[62385]: DEBUG nova.virt.hardware [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 990.359736] env[62385]: DEBUG nova.virt.hardware [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 990.360154] env[62385]: DEBUG nova.virt.hardware [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 990.360492] env[62385]: DEBUG nova.virt.hardware [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 990.367420] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Reconfiguring VM instance instance-00000061 to detach disk 2000 {{(pid=62385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 990.368429] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f24adb9c-b53a-4ce7-a8c0-ec6effb0acbe {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.395554] env[62385]: DEBUG oslo_vmware.api [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 990.395554] env[62385]: value = "task-1206263" [ 990.395554] env[62385]: _type = "Task" [ 990.395554] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.405055] env[62385]: DEBUG oslo_vmware.api [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206263, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.450495] env[62385]: DEBUG nova.compute.manager [req-18412fb3-0cfa-4076-95e9-c6f0351b6b7c req-78cdcdda-aaaa-40ec-858a-d9e68db92f64 service nova] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Received event network-vif-deleted-a755329d-e122-40af-add4-88f11f88402c {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 990.450689] env[62385]: INFO nova.compute.manager [req-18412fb3-0cfa-4076-95e9-c6f0351b6b7c req-78cdcdda-aaaa-40ec-858a-d9e68db92f64 service nova] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Neutron deleted interface a755329d-e122-40af-add4-88f11f88402c; detaching it from the instance and deleting it from the info cache [ 990.450771] env[62385]: DEBUG nova.network.neutron [req-18412fb3-0cfa-4076-95e9-c6f0351b6b7c req-78cdcdda-aaaa-40ec-858a-d9e68db92f64 service nova] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.523601] env[62385]: DEBUG nova.scheduler.client.report [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 990.558118] env[62385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 990.558118] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e4ff39-fc59-5b58-ec9e-cb3a9fb0730c" [ 990.558118] env[62385]: _type = "HttpNfcLease" [ 990.558118] env[62385]: } is ready. {{(pid=62385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 990.558448] env[62385]: DEBUG oslo_vmware.rw_handles [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 990.558448] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e4ff39-fc59-5b58-ec9e-cb3a9fb0730c" [ 990.558448] env[62385]: _type = "HttpNfcLease" [ 990.558448] env[62385]: }. {{(pid=62385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 990.559172] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abea026a-7f2c-40d3-b18d-a1297939a570 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.571411] env[62385]: DEBUG oslo_vmware.rw_handles [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52515341-9ab6-3170-21a5-eb2804418e3d/disk-0.vmdk from lease info. {{(pid=62385) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 990.571596] env[62385]: DEBUG oslo_vmware.rw_handles [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52515341-9ab6-3170-21a5-eb2804418e3d/disk-0.vmdk for reading. {{(pid=62385) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 990.637421] env[62385]: DEBUG nova.network.neutron [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Updating instance_info_cache with network_info: [{"id": "a9513285-4023-4f3e-b90d-3eb507955f13", "address": "fa:16:3e:30:02:c0", "network": {"id": "30c4e78b-6a64-47dc-aaaa-a210160c9f23", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1809587656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "105150c170d74e268e82ab894e9bf0d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10ff2092-e8eb-4768-ad4a-65a80560b447", "external-id": "nsx-vlan-transportzone-845", "segmentation_id": 845, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa9513285-40", "ovs_interfaceid": "a9513285-4023-4f3e-b90d-3eb507955f13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.683853] env[62385]: DEBUG nova.compute.manager [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 990.706424] env[62385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-0297668c-98ab-4550-8a5c-811982570f27 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.711310] env[62385]: DEBUG nova.virt.hardware [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 990.711591] env[62385]: DEBUG nova.virt.hardware [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 990.711833] env[62385]: DEBUG nova.virt.hardware [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 990.712096] env[62385]: DEBUG nova.virt.hardware [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 990.712334] env[62385]: DEBUG nova.virt.hardware [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 990.712541] env[62385]: DEBUG nova.virt.hardware [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 990.712814] env[62385]: DEBUG nova.virt.hardware [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 990.713201] env[62385]: DEBUG nova.virt.hardware [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 990.713337] env[62385]: DEBUG nova.virt.hardware [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 990.713639] env[62385]: DEBUG nova.virt.hardware [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 990.713856] env[62385]: DEBUG nova.virt.hardware [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 990.715103] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4f47381-3b7c-46a2-ad84-1805a795c14e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.729113] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e39ae86e-d7d5-4b06-ac6a-4e0482a9df6f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.910210] env[62385]: DEBUG oslo_vmware.api [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206263, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.919752] env[62385]: DEBUG nova.network.neutron [-] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.957548] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-df42a6a6-a158-4f3f-9ec9-0cb63a9c233c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.969987] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98b91b9a-ad7f-4467-acef-73649cebdeb9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.003100] env[62385]: DEBUG nova.compute.manager [req-18412fb3-0cfa-4076-95e9-c6f0351b6b7c req-78cdcdda-aaaa-40ec-858a-d9e68db92f64 service nova] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Detach interface failed, port_id=a755329d-e122-40af-add4-88f11f88402c, reason: Instance 9741e9c8-60d9-41bf-8185-ef3cfcc1c844 could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 991.007481] env[62385]: DEBUG nova.compute.manager [req-17b38db0-4923-4e80-9de4-ce0cc79583b1 req-f40a9edc-7aa6-48ba-ab24-1cbd8a45cac7 service nova] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Received event network-vif-plugged-117fbbdf-9bdd-4d68-ac9f-d2fa6279aebd {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 991.007481] env[62385]: DEBUG oslo_concurrency.lockutils [req-17b38db0-4923-4e80-9de4-ce0cc79583b1 req-f40a9edc-7aa6-48ba-ab24-1cbd8a45cac7 service nova] Acquiring lock "76672404-2bcf-438a-9e21-92ea9dc86461-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.007481] env[62385]: DEBUG oslo_concurrency.lockutils [req-17b38db0-4923-4e80-9de4-ce0cc79583b1 req-f40a9edc-7aa6-48ba-ab24-1cbd8a45cac7 service nova] Lock "76672404-2bcf-438a-9e21-92ea9dc86461-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.007863] env[62385]: DEBUG oslo_concurrency.lockutils [req-17b38db0-4923-4e80-9de4-ce0cc79583b1 req-f40a9edc-7aa6-48ba-ab24-1cbd8a45cac7 service nova] Lock "76672404-2bcf-438a-9e21-92ea9dc86461-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.008014] env[62385]: DEBUG nova.compute.manager [req-17b38db0-4923-4e80-9de4-ce0cc79583b1 req-f40a9edc-7aa6-48ba-ab24-1cbd8a45cac7 service nova] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] No waiting events found dispatching network-vif-plugged-117fbbdf-9bdd-4d68-ac9f-d2fa6279aebd {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 991.008257] env[62385]: WARNING nova.compute.manager [req-17b38db0-4923-4e80-9de4-ce0cc79583b1 req-f40a9edc-7aa6-48ba-ab24-1cbd8a45cac7 service nova] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Received unexpected event network-vif-plugged-117fbbdf-9bdd-4d68-ac9f-d2fa6279aebd for instance with vm_state building and task_state spawning. [ 991.029873] env[62385]: DEBUG oslo_concurrency.lockutils [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.375s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.030786] env[62385]: DEBUG nova.compute.manager [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] [instance: 784589bb-a668-4674-83b8-6219def15067] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 991.034722] env[62385]: DEBUG oslo_concurrency.lockutils [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.930s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.036332] env[62385]: INFO nova.compute.claims [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 991.140144] env[62385]: DEBUG oslo_concurrency.lockutils [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Releasing lock "refresh_cache-df020d6b-3fab-4599-a342-47c7833b4240" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.154110] env[62385]: DEBUG nova.network.neutron [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Successfully updated port: 117fbbdf-9bdd-4d68-ac9f-d2fa6279aebd {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 991.408414] env[62385]: DEBUG oslo_vmware.api [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206263, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.422212] env[62385]: INFO nova.compute.manager [-] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Took 1.40 seconds to deallocate network for instance. [ 991.547060] env[62385]: DEBUG nova.compute.utils [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 991.550842] env[62385]: DEBUG nova.compute.manager [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] [instance: 784589bb-a668-4674-83b8-6219def15067] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 991.554322] env[62385]: DEBUG nova.network.neutron [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] [instance: 784589bb-a668-4674-83b8-6219def15067] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 991.621349] env[62385]: DEBUG nova.policy [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '21a855054740402cbb3dfbf623aa4525', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e1bce5a5291b45fcb766cc6b694e45c6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 991.657216] env[62385]: DEBUG oslo_concurrency.lockutils [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "refresh_cache-76672404-2bcf-438a-9e21-92ea9dc86461" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.657413] env[62385]: DEBUG oslo_concurrency.lockutils [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquired lock "refresh_cache-76672404-2bcf-438a-9e21-92ea9dc86461" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.657537] env[62385]: DEBUG nova.network.neutron [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 991.681152] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21251196-a2e7-4d7e-ab55-c8ac128e8400 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.702918] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95629d7d-f35e-48f7-accf-35614afdbc8a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.712029] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Updating instance 'df020d6b-3fab-4599-a342-47c7833b4240' progress to 83 {{(pid=62385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 991.912152] env[62385]: DEBUG oslo_vmware.api [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206263, 'name': ReconfigVM_Task, 'duration_secs': 1.198239} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.912152] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Reconfigured VM instance instance-00000061 to detach disk 2000 {{(pid=62385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 991.912152] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05343037-6467-4d0b-b653-67dfdd71fbf6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.934742] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a3a6fc1a-c0ca-4fe5-9a14-65b9435d0e5f tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.946727] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] b1e0baf3-643c-49c2-8a80-8de07d8527e7/b1e0baf3-643c-49c2-8a80-8de07d8527e7.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 991.947276] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5dba2b66-cb6e-452e-b93c-610f9500de0c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.963347] env[62385]: DEBUG nova.network.neutron [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] [instance: 784589bb-a668-4674-83b8-6219def15067] Successfully created port: 682074de-7041-442e-bc00-fcefe591c878 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 991.971092] env[62385]: DEBUG oslo_vmware.api [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 991.971092] env[62385]: value = "task-1206264" [ 991.971092] env[62385]: _type = "Task" [ 991.971092] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.979390] env[62385]: DEBUG oslo_vmware.api [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206264, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.058507] env[62385]: DEBUG nova.compute.manager [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] [instance: 784589bb-a668-4674-83b8-6219def15067] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 992.218940] env[62385]: DEBUG nova.network.neutron [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 992.222906] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 992.223056] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-303173fd-8a04-43c3-b4ed-4da1065ad5b1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.234748] env[62385]: DEBUG oslo_vmware.api [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 992.234748] env[62385]: value = "task-1206265" [ 992.234748] env[62385]: _type = "Task" [ 992.234748] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.245757] env[62385]: DEBUG oslo_vmware.api [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206265, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.337644] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a46b65c-e865-4bea-bc68-2be6bca1c4bb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.346892] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da805991-62c7-4a15-b9a1-c3a0ad2430c6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.381777] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-475dd21f-b70a-4760-b6f2-daaab99e9e49 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.389951] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23295538-211d-4b01-81dc-f10f94673a3a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.406991] env[62385]: DEBUG nova.compute.provider_tree [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 992.484670] env[62385]: DEBUG oslo_vmware.api [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206264, 'name': ReconfigVM_Task, 'duration_secs': 0.29069} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.485067] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Reconfigured VM instance instance-00000061 to attach disk [datastore2] b1e0baf3-643c-49c2-8a80-8de07d8527e7/b1e0baf3-643c-49c2-8a80-8de07d8527e7.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 992.485397] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Updating instance 'b1e0baf3-643c-49c2-8a80-8de07d8527e7' progress to 50 {{(pid=62385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 992.539085] env[62385]: DEBUG nova.network.neutron [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Updating instance_info_cache with network_info: [{"id": "117fbbdf-9bdd-4d68-ac9f-d2fa6279aebd", "address": "fa:16:3e:27:4c:11", "network": {"id": "8b2d7a6e-25d4-4a54-ab54-5972612046da", "bridge": "br-int", "label": "tempest-ServersTestJSON-1142134211-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bf8e3c942d3445919cfbe988cca84e90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap117fbbdf-9b", "ovs_interfaceid": "117fbbdf-9bdd-4d68-ac9f-d2fa6279aebd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.743699] env[62385]: DEBUG oslo_vmware.api [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206265, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.910781] env[62385]: DEBUG nova.scheduler.client.report [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 992.993792] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7efdb39b-0bad-4a5d-b88e-0d137a911fad {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.015033] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b295294-944c-4ff0-b4b6-a7915da510dd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.034137] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Updating instance 'b1e0baf3-643c-49c2-8a80-8de07d8527e7' progress to 67 {{(pid=62385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 993.039656] env[62385]: DEBUG nova.compute.manager [req-8d5d1410-6447-4c25-af3d-449c2f16f506 req-0ea635d5-96c2-472a-b784-ed7cc26c2115 service nova] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Received event network-changed-117fbbdf-9bdd-4d68-ac9f-d2fa6279aebd {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 993.039874] env[62385]: DEBUG nova.compute.manager [req-8d5d1410-6447-4c25-af3d-449c2f16f506 req-0ea635d5-96c2-472a-b784-ed7cc26c2115 service nova] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Refreshing instance network info cache due to event network-changed-117fbbdf-9bdd-4d68-ac9f-d2fa6279aebd. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 993.040101] env[62385]: DEBUG oslo_concurrency.lockutils [req-8d5d1410-6447-4c25-af3d-449c2f16f506 req-0ea635d5-96c2-472a-b784-ed7cc26c2115 service nova] Acquiring lock "refresh_cache-76672404-2bcf-438a-9e21-92ea9dc86461" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.042022] env[62385]: DEBUG oslo_concurrency.lockutils [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Releasing lock "refresh_cache-76672404-2bcf-438a-9e21-92ea9dc86461" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.042316] env[62385]: DEBUG nova.compute.manager [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Instance network_info: |[{"id": "117fbbdf-9bdd-4d68-ac9f-d2fa6279aebd", "address": "fa:16:3e:27:4c:11", "network": {"id": "8b2d7a6e-25d4-4a54-ab54-5972612046da", "bridge": "br-int", "label": "tempest-ServersTestJSON-1142134211-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bf8e3c942d3445919cfbe988cca84e90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap117fbbdf-9b", "ovs_interfaceid": "117fbbdf-9bdd-4d68-ac9f-d2fa6279aebd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 993.042585] env[62385]: DEBUG oslo_concurrency.lockutils [req-8d5d1410-6447-4c25-af3d-449c2f16f506 req-0ea635d5-96c2-472a-b784-ed7cc26c2115 service nova] Acquired lock "refresh_cache-76672404-2bcf-438a-9e21-92ea9dc86461" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.042764] env[62385]: DEBUG nova.network.neutron [req-8d5d1410-6447-4c25-af3d-449c2f16f506 req-0ea635d5-96c2-472a-b784-ed7cc26c2115 service nova] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Refreshing network info cache for port 117fbbdf-9bdd-4d68-ac9f-d2fa6279aebd {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 993.043926] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:27:4c:11', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '117fbbdf-9bdd-4d68-ac9f-d2fa6279aebd', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 993.052634] env[62385]: DEBUG oslo.service.loopingcall [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 993.053477] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 993.053735] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2e357f47-de29-44fc-b19e-8d9f7bd7ef01 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.070765] env[62385]: DEBUG nova.compute.manager [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] [instance: 784589bb-a668-4674-83b8-6219def15067] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 993.079623] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 993.079623] env[62385]: value = "task-1206269" [ 993.079623] env[62385]: _type = "Task" [ 993.079623] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.092053] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206269, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.094659] env[62385]: DEBUG nova.virt.hardware [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 993.095027] env[62385]: DEBUG nova.virt.hardware [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 993.095247] env[62385]: DEBUG nova.virt.hardware [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 993.095495] env[62385]: DEBUG nova.virt.hardware [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 993.095680] env[62385]: DEBUG nova.virt.hardware [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 993.095908] env[62385]: DEBUG nova.virt.hardware [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 993.096238] env[62385]: DEBUG nova.virt.hardware [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 993.096455] env[62385]: DEBUG nova.virt.hardware [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 993.096688] env[62385]: DEBUG nova.virt.hardware [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 993.096909] env[62385]: DEBUG nova.virt.hardware [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 993.097168] env[62385]: DEBUG nova.virt.hardware [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 993.098684] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dc66da7-377d-4188-b8d9-1ea24408b905 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.108073] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82a7d70a-05c9-4890-a238-914947b848d6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.245168] env[62385]: DEBUG oslo_vmware.api [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206265, 'name': PowerOnVM_Task, 'duration_secs': 0.581084} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.245516] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 993.245814] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-565e98b2-825e-442f-b984-060fc8525aaa tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Updating instance 'df020d6b-3fab-4599-a342-47c7833b4240' progress to 100 {{(pid=62385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 993.418032] env[62385]: DEBUG oslo_concurrency.lockutils [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.382s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.418032] env[62385]: DEBUG nova.compute.manager [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 993.420479] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 3.672s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.514125] env[62385]: DEBUG nova.network.neutron [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] [instance: 784589bb-a668-4674-83b8-6219def15067] Successfully updated port: 682074de-7041-442e-bc00-fcefe591c878 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 993.589840] env[62385]: DEBUG nova.network.neutron [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Port eef524f4-fc00-47f1-a485-dcf568c6d0e4 binding to destination host cpu-1 is already ACTIVE {{(pid=62385) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 993.597607] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206269, 'name': CreateVM_Task, 'duration_secs': 0.387745} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.598410] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 993.598820] env[62385]: DEBUG oslo_concurrency.lockutils [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.599017] env[62385]: DEBUG oslo_concurrency.lockutils [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.599345] env[62385]: DEBUG oslo_concurrency.lockutils [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 993.599611] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-09f29e01-a3ce-4e23-b8d7-3d8c78787a62 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.605788] env[62385]: DEBUG oslo_vmware.api [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 993.605788] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5210c4cb-c05b-5baa-abc8-3f996fac4585" [ 993.605788] env[62385]: _type = "Task" [ 993.605788] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.620026] env[62385]: DEBUG oslo_vmware.api [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5210c4cb-c05b-5baa-abc8-3f996fac4585, 'name': SearchDatastore_Task, 'duration_secs': 0.01033} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.620347] env[62385]: DEBUG oslo_concurrency.lockutils [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.620611] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 993.622172] env[62385]: DEBUG oslo_concurrency.lockutils [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.622172] env[62385]: DEBUG oslo_concurrency.lockutils [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.622172] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 993.622172] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2dd6b849-c5bb-4e0e-8284-2a9ada5025d5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.629317] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 993.629508] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 993.630404] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bedf4e2f-d2d0-483f-a207-275389d21c79 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.635869] env[62385]: DEBUG oslo_vmware.api [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 993.635869] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a21f4e-2f4e-0b0d-e244-c64cddbf2e10" [ 993.635869] env[62385]: _type = "Task" [ 993.635869] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.643433] env[62385]: DEBUG oslo_vmware.api [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a21f4e-2f4e-0b0d-e244-c64cddbf2e10, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.923928] env[62385]: DEBUG nova.compute.utils [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 993.928125] env[62385]: INFO nova.compute.claims [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 993.933052] env[62385]: DEBUG nova.compute.manager [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 993.933225] env[62385]: DEBUG nova.network.neutron [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 993.944028] env[62385]: DEBUG nova.network.neutron [req-8d5d1410-6447-4c25-af3d-449c2f16f506 req-0ea635d5-96c2-472a-b784-ed7cc26c2115 service nova] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Updated VIF entry in instance network info cache for port 117fbbdf-9bdd-4d68-ac9f-d2fa6279aebd. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 993.944028] env[62385]: DEBUG nova.network.neutron [req-8d5d1410-6447-4c25-af3d-449c2f16f506 req-0ea635d5-96c2-472a-b784-ed7cc26c2115 service nova] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Updating instance_info_cache with network_info: [{"id": "117fbbdf-9bdd-4d68-ac9f-d2fa6279aebd", "address": "fa:16:3e:27:4c:11", "network": {"id": "8b2d7a6e-25d4-4a54-ab54-5972612046da", "bridge": "br-int", "label": "tempest-ServersTestJSON-1142134211-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bf8e3c942d3445919cfbe988cca84e90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap117fbbdf-9b", "ovs_interfaceid": "117fbbdf-9bdd-4d68-ac9f-d2fa6279aebd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.976513] env[62385]: DEBUG nova.policy [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5040b20c848c4d28a198f1773a3ff21e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '012426f098ce40c3aaa00f628fe9cebb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 994.016402] env[62385]: DEBUG oslo_concurrency.lockutils [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Acquiring lock "refresh_cache-784589bb-a668-4674-83b8-6219def15067" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.016578] env[62385]: DEBUG oslo_concurrency.lockutils [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Acquired lock "refresh_cache-784589bb-a668-4674-83b8-6219def15067" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.016711] env[62385]: DEBUG nova.network.neutron [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] [instance: 784589bb-a668-4674-83b8-6219def15067] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 994.148910] env[62385]: DEBUG oslo_vmware.api [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a21f4e-2f4e-0b0d-e244-c64cddbf2e10, 'name': SearchDatastore_Task, 'duration_secs': 0.009914} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.150462] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf725352-9e8a-4a87-91c2-972aa64dcd49 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.156432] env[62385]: DEBUG oslo_vmware.api [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 994.156432] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c2a009-0928-1aa6-b393-68440cd52623" [ 994.156432] env[62385]: _type = "Task" [ 994.156432] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.164493] env[62385]: DEBUG oslo_vmware.api [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c2a009-0928-1aa6-b393-68440cd52623, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.280354] env[62385]: DEBUG nova.network.neutron [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Successfully created port: a7153a2d-db69-4d14-8ddc-b634ddabbc18 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 994.433830] env[62385]: DEBUG nova.compute.manager [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 994.440069] env[62385]: INFO nova.compute.resource_tracker [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Updating resource usage from migration a29e1d61-503d-4be1-9d07-d708bf2b943e [ 994.445377] env[62385]: DEBUG oslo_concurrency.lockutils [req-8d5d1410-6447-4c25-af3d-449c2f16f506 req-0ea635d5-96c2-472a-b784-ed7cc26c2115 service nova] Releasing lock "refresh_cache-76672404-2bcf-438a-9e21-92ea9dc86461" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.552184] env[62385]: DEBUG nova.network.neutron [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] [instance: 784589bb-a668-4674-83b8-6219def15067] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 994.615184] env[62385]: DEBUG oslo_concurrency.lockutils [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "b1e0baf3-643c-49c2-8a80-8de07d8527e7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.615480] env[62385]: DEBUG oslo_concurrency.lockutils [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "b1e0baf3-643c-49c2-8a80-8de07d8527e7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.615675] env[62385]: DEBUG oslo_concurrency.lockutils [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "b1e0baf3-643c-49c2-8a80-8de07d8527e7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.632816] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e8df002-e2a6-4724-9449-ed3b676347bc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.642111] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-801927d5-3959-4a26-ba4c-be98f6f0e336 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.678337] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf5ac97d-7c56-4806-a4b3-13c0fc2daa8e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.689834] env[62385]: DEBUG oslo_vmware.api [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c2a009-0928-1aa6-b393-68440cd52623, 'name': SearchDatastore_Task, 'duration_secs': 0.009811} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.690494] env[62385]: DEBUG oslo_concurrency.lockutils [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.690771] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 76672404-2bcf-438a-9e21-92ea9dc86461/76672404-2bcf-438a-9e21-92ea9dc86461.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 994.692198] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-684b343f-17ee-43d0-b85b-9b0c74bc5c81 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.696190] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a3bb88b7-f5d5-4b85-99b7-650be4d3eae0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.708565] env[62385]: DEBUG nova.compute.provider_tree [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 994.712843] env[62385]: DEBUG oslo_vmware.api [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 994.712843] env[62385]: value = "task-1206270" [ 994.712843] env[62385]: _type = "Task" [ 994.712843] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.721364] env[62385]: DEBUG oslo_vmware.api [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206270, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.046660] env[62385]: DEBUG nova.network.neutron [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] [instance: 784589bb-a668-4674-83b8-6219def15067] Updating instance_info_cache with network_info: [{"id": "682074de-7041-442e-bc00-fcefe591c878", "address": "fa:16:3e:c1:3f:33", "network": {"id": "e9808126-67bd-4f8e-9314-e398fc95f31d", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-847554994-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1bce5a5291b45fcb766cc6b694e45c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8868dc2-7767-49c0-a2ed-e611fcbf8414", "external-id": "nsx-vlan-transportzone-158", "segmentation_id": 158, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap682074de-70", "ovs_interfaceid": "682074de-7041-442e-bc00-fcefe591c878", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.215258] env[62385]: DEBUG nova.scheduler.client.report [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 995.237218] env[62385]: DEBUG oslo_vmware.api [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206270, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.513667} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.237439] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 76672404-2bcf-438a-9e21-92ea9dc86461/76672404-2bcf-438a-9e21-92ea9dc86461.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 995.237675] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 995.237952] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d7ae5555-841f-4a09-8ceb-97033dfbc0eb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.247930] env[62385]: DEBUG oslo_vmware.api [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 995.247930] env[62385]: value = "task-1206271" [ 995.247930] env[62385]: _type = "Task" [ 995.247930] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.256988] env[62385]: DEBUG oslo_vmware.api [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206271, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.418677] env[62385]: DEBUG nova.compute.manager [req-b5e7d32f-a810-45dd-8a6b-3be0698e4b48 req-83a6d099-a3b5-40e3-b3bd-4d58287edffa service nova] [instance: 784589bb-a668-4674-83b8-6219def15067] Received event network-vif-plugged-682074de-7041-442e-bc00-fcefe591c878 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 995.418903] env[62385]: DEBUG oslo_concurrency.lockutils [req-b5e7d32f-a810-45dd-8a6b-3be0698e4b48 req-83a6d099-a3b5-40e3-b3bd-4d58287edffa service nova] Acquiring lock "784589bb-a668-4674-83b8-6219def15067-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.419133] env[62385]: DEBUG oslo_concurrency.lockutils [req-b5e7d32f-a810-45dd-8a6b-3be0698e4b48 req-83a6d099-a3b5-40e3-b3bd-4d58287edffa service nova] Lock "784589bb-a668-4674-83b8-6219def15067-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.419308] env[62385]: DEBUG oslo_concurrency.lockutils [req-b5e7d32f-a810-45dd-8a6b-3be0698e4b48 req-83a6d099-a3b5-40e3-b3bd-4d58287edffa service nova] Lock "784589bb-a668-4674-83b8-6219def15067-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.419478] env[62385]: DEBUG nova.compute.manager [req-b5e7d32f-a810-45dd-8a6b-3be0698e4b48 req-83a6d099-a3b5-40e3-b3bd-4d58287edffa service nova] [instance: 784589bb-a668-4674-83b8-6219def15067] No waiting events found dispatching network-vif-plugged-682074de-7041-442e-bc00-fcefe591c878 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 995.419642] env[62385]: WARNING nova.compute.manager [req-b5e7d32f-a810-45dd-8a6b-3be0698e4b48 req-83a6d099-a3b5-40e3-b3bd-4d58287edffa service nova] [instance: 784589bb-a668-4674-83b8-6219def15067] Received unexpected event network-vif-plugged-682074de-7041-442e-bc00-fcefe591c878 for instance with vm_state building and task_state spawning. [ 995.419819] env[62385]: DEBUG nova.compute.manager [req-b5e7d32f-a810-45dd-8a6b-3be0698e4b48 req-83a6d099-a3b5-40e3-b3bd-4d58287edffa service nova] [instance: 784589bb-a668-4674-83b8-6219def15067] Received event network-changed-682074de-7041-442e-bc00-fcefe591c878 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 995.419978] env[62385]: DEBUG nova.compute.manager [req-b5e7d32f-a810-45dd-8a6b-3be0698e4b48 req-83a6d099-a3b5-40e3-b3bd-4d58287edffa service nova] [instance: 784589bb-a668-4674-83b8-6219def15067] Refreshing instance network info cache due to event network-changed-682074de-7041-442e-bc00-fcefe591c878. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 995.420187] env[62385]: DEBUG oslo_concurrency.lockutils [req-b5e7d32f-a810-45dd-8a6b-3be0698e4b48 req-83a6d099-a3b5-40e3-b3bd-4d58287edffa service nova] Acquiring lock "refresh_cache-784589bb-a668-4674-83b8-6219def15067" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.444938] env[62385]: DEBUG nova.compute.manager [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 995.467072] env[62385]: DEBUG nova.virt.hardware [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 995.467345] env[62385]: DEBUG nova.virt.hardware [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 995.467511] env[62385]: DEBUG nova.virt.hardware [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 995.467699] env[62385]: DEBUG nova.virt.hardware [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 995.467855] env[62385]: DEBUG nova.virt.hardware [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 995.468017] env[62385]: DEBUG nova.virt.hardware [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 995.468245] env[62385]: DEBUG nova.virt.hardware [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 995.468413] env[62385]: DEBUG nova.virt.hardware [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 995.468586] env[62385]: DEBUG nova.virt.hardware [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 995.468752] env[62385]: DEBUG nova.virt.hardware [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 995.468932] env[62385]: DEBUG nova.virt.hardware [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 995.469851] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-773269f0-94d1-4812-87e0-617d094ab514 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.479023] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e417b028-7062-45df-a5f9-f346ab9d60a9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.552563] env[62385]: DEBUG oslo_concurrency.lockutils [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Releasing lock "refresh_cache-784589bb-a668-4674-83b8-6219def15067" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.552563] env[62385]: DEBUG nova.compute.manager [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] [instance: 784589bb-a668-4674-83b8-6219def15067] Instance network_info: |[{"id": "682074de-7041-442e-bc00-fcefe591c878", "address": "fa:16:3e:c1:3f:33", "network": {"id": "e9808126-67bd-4f8e-9314-e398fc95f31d", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-847554994-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1bce5a5291b45fcb766cc6b694e45c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8868dc2-7767-49c0-a2ed-e611fcbf8414", "external-id": "nsx-vlan-transportzone-158", "segmentation_id": 158, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap682074de-70", "ovs_interfaceid": "682074de-7041-442e-bc00-fcefe591c878", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 995.552563] env[62385]: DEBUG oslo_concurrency.lockutils [req-b5e7d32f-a810-45dd-8a6b-3be0698e4b48 req-83a6d099-a3b5-40e3-b3bd-4d58287edffa service nova] Acquired lock "refresh_cache-784589bb-a668-4674-83b8-6219def15067" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.552563] env[62385]: DEBUG nova.network.neutron [req-b5e7d32f-a810-45dd-8a6b-3be0698e4b48 req-83a6d099-a3b5-40e3-b3bd-4d58287edffa service nova] [instance: 784589bb-a668-4674-83b8-6219def15067] Refreshing network info cache for port 682074de-7041-442e-bc00-fcefe591c878 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 995.553291] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] [instance: 784589bb-a668-4674-83b8-6219def15067] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c1:3f:33', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c8868dc2-7767-49c0-a2ed-e611fcbf8414', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '682074de-7041-442e-bc00-fcefe591c878', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 995.561955] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Creating folder: Project (e1bce5a5291b45fcb766cc6b694e45c6). Parent ref: group-v261107. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 995.563300] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e6ab1b07-2966-4e37-9787-e7621c272dd2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.574677] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Created folder: Project (e1bce5a5291b45fcb766cc6b694e45c6) in parent group-v261107. [ 995.574677] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Creating folder: Instances. Parent ref: group-v261282. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 995.574991] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-49e26384-503d-4f04-abad-85887da4b8b3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.584169] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Created folder: Instances in parent group-v261282. [ 995.584215] env[62385]: DEBUG oslo.service.loopingcall [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 995.584408] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 784589bb-a668-4674-83b8-6219def15067] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 995.584622] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b8119f2a-fdb3-471c-9d41-972808cc1806 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.606629] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 995.606629] env[62385]: value = "task-1206274" [ 995.606629] env[62385]: _type = "Task" [ 995.606629] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.615871] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206274, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.700160] env[62385]: DEBUG oslo_concurrency.lockutils [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "refresh_cache-b1e0baf3-643c-49c2-8a80-8de07d8527e7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.700160] env[62385]: DEBUG oslo_concurrency.lockutils [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquired lock "refresh_cache-b1e0baf3-643c-49c2-8a80-8de07d8527e7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.700160] env[62385]: DEBUG nova.network.neutron [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 995.721182] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.300s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.721182] env[62385]: INFO nova.compute.manager [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Migrating [ 995.728741] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a3a6fc1a-c0ca-4fe5-9a14-65b9435d0e5f tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.795s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.729074] env[62385]: DEBUG nova.objects.instance [None req-a3a6fc1a-c0ca-4fe5-9a14-65b9435d0e5f tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lazy-loading 'resources' on Instance uuid 9741e9c8-60d9-41bf-8185-ef3cfcc1c844 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 995.767089] env[62385]: DEBUG oslo_vmware.api [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206271, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.090741} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.767089] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 995.767420] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9dbd85f-6bd6-4bc3-ac2c-5b3b2aad41a3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.803889] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] 76672404-2bcf-438a-9e21-92ea9dc86461/76672404-2bcf-438a-9e21-92ea9dc86461.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 995.805291] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-835ee4c0-13a6-4b80-9393-67804c5c57c1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.824058] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9eff707c-16e0-46c5-b85d-791396fd0db5 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "5daf0a99-0c2a-40d8-afc7-1998e21e32d8" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.824334] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9eff707c-16e0-46c5-b85d-791396fd0db5 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "5daf0a99-0c2a-40d8-afc7-1998e21e32d8" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.829183] env[62385]: DEBUG oslo_vmware.api [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 995.829183] env[62385]: value = "task-1206275" [ 995.829183] env[62385]: _type = "Task" [ 995.829183] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.840721] env[62385]: DEBUG oslo_vmware.api [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206275, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.843097] env[62385]: DEBUG nova.network.neutron [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Successfully updated port: a7153a2d-db69-4d14-8ddc-b634ddabbc18 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 996.031538] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8e10c3aa-33db-4d4b-ae34-41f4b8a0ebec tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "df020d6b-3fab-4599-a342-47c7833b4240" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.031852] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8e10c3aa-33db-4d4b-ae34-41f4b8a0ebec tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "df020d6b-3fab-4599-a342-47c7833b4240" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.032076] env[62385]: DEBUG nova.compute.manager [None req-8e10c3aa-33db-4d4b-ae34-41f4b8a0ebec tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Going to confirm migration 2 {{(pid=62385) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 996.118025] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206274, 'name': CreateVM_Task, 'duration_secs': 0.493625} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.118025] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 784589bb-a668-4674-83b8-6219def15067] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 996.118224] env[62385]: DEBUG oslo_concurrency.lockutils [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.118255] env[62385]: DEBUG oslo_concurrency.lockutils [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.118576] env[62385]: DEBUG oslo_concurrency.lockutils [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 996.118962] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b9ff234a-82ef-4e21-ab35-c3131e01a390 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.124909] env[62385]: DEBUG oslo_vmware.api [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Waiting for the task: (returnval){ [ 996.124909] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e268af-5abf-94f1-5934-217dc42dcc3c" [ 996.124909] env[62385]: _type = "Task" [ 996.124909] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.135769] env[62385]: DEBUG oslo_vmware.api [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e268af-5abf-94f1-5934-217dc42dcc3c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.241446] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "refresh_cache-9b9fd167-d843-4b9d-9e4f-dbde5d3d629e" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.242472] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquired lock "refresh_cache-9b9fd167-d843-4b9d-9e4f-dbde5d3d629e" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.242472] env[62385]: DEBUG nova.network.neutron [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 996.295941] env[62385]: DEBUG nova.network.neutron [req-b5e7d32f-a810-45dd-8a6b-3be0698e4b48 req-83a6d099-a3b5-40e3-b3bd-4d58287edffa service nova] [instance: 784589bb-a668-4674-83b8-6219def15067] Updated VIF entry in instance network info cache for port 682074de-7041-442e-bc00-fcefe591c878. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 996.296374] env[62385]: DEBUG nova.network.neutron [req-b5e7d32f-a810-45dd-8a6b-3be0698e4b48 req-83a6d099-a3b5-40e3-b3bd-4d58287edffa service nova] [instance: 784589bb-a668-4674-83b8-6219def15067] Updating instance_info_cache with network_info: [{"id": "682074de-7041-442e-bc00-fcefe591c878", "address": "fa:16:3e:c1:3f:33", "network": {"id": "e9808126-67bd-4f8e-9314-e398fc95f31d", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-847554994-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1bce5a5291b45fcb766cc6b694e45c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8868dc2-7767-49c0-a2ed-e611fcbf8414", "external-id": "nsx-vlan-transportzone-158", "segmentation_id": 158, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap682074de-70", "ovs_interfaceid": "682074de-7041-442e-bc00-fcefe591c878", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.328555] env[62385]: DEBUG nova.compute.utils [None req-9eff707c-16e0-46c5-b85d-791396fd0db5 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 996.342557] env[62385]: DEBUG oslo_vmware.api [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206275, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.347731] env[62385]: DEBUG oslo_concurrency.lockutils [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "refresh_cache-57637d6e-39e2-405c-bbcf-64bf0ffbebd1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.347898] env[62385]: DEBUG oslo_concurrency.lockutils [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquired lock "refresh_cache-57637d6e-39e2-405c-bbcf-64bf0ffbebd1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.348266] env[62385]: DEBUG nova.network.neutron [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 996.477380] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a8d6d9f-3da0-405e-8cdc-22cea1629516 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.486434] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a9afd4a-68ff-4404-83c6-5a28fcf57fd3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.526008] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c863e44-8615-4c54-9cb9-663f77a54458 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.540752] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-343b75b9-c511-4559-bbca-24b2edbc89b6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.558303] env[62385]: DEBUG nova.compute.provider_tree [None req-a3a6fc1a-c0ca-4fe5-9a14-65b9435d0e5f tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 996.595146] env[62385]: DEBUG nova.network.neutron [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Updating instance_info_cache with network_info: [{"id": "eef524f4-fc00-47f1-a485-dcf568c6d0e4", "address": "fa:16:3e:e6:2c:55", "network": {"id": "ff21dd88-fa02-4bd6-8009-980ec6e72758", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1065789169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aea48968d4464ac0ab01c998c059107d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3ff3baee-99ce-4b51-ae98-efc6163aaab3", "external-id": "nsx-vlan-transportzone-574", "segmentation_id": 574, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeef524f4-fc", "ovs_interfaceid": "eef524f4-fc00-47f1-a485-dcf568c6d0e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.609017] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8e10c3aa-33db-4d4b-ae34-41f4b8a0ebec tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "refresh_cache-df020d6b-3fab-4599-a342-47c7833b4240" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.609216] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8e10c3aa-33db-4d4b-ae34-41f4b8a0ebec tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquired lock "refresh_cache-df020d6b-3fab-4599-a342-47c7833b4240" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.609395] env[62385]: DEBUG nova.network.neutron [None req-8e10c3aa-33db-4d4b-ae34-41f4b8a0ebec tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 996.609577] env[62385]: DEBUG nova.objects.instance [None req-8e10c3aa-33db-4d4b-ae34-41f4b8a0ebec tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lazy-loading 'info_cache' on Instance uuid df020d6b-3fab-4599-a342-47c7833b4240 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 996.636305] env[62385]: DEBUG oslo_vmware.api [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e268af-5abf-94f1-5934-217dc42dcc3c, 'name': SearchDatastore_Task, 'duration_secs': 0.034772} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.636601] env[62385]: DEBUG oslo_concurrency.lockutils [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.636916] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] [instance: 784589bb-a668-4674-83b8-6219def15067] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 996.637122] env[62385]: DEBUG oslo_concurrency.lockutils [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.637280] env[62385]: DEBUG oslo_concurrency.lockutils [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.637463] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 996.637826] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e09d5d52-21c0-41ed-a3ce-2e3acbf84c9b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.646394] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 996.646579] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 996.647313] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-45ee65bb-bf85-4987-8bf0-e85a0339afa3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.652931] env[62385]: DEBUG oslo_vmware.api [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Waiting for the task: (returnval){ [ 996.652931] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5234f616-8fab-3ae3-54c4-53469f3acb39" [ 996.652931] env[62385]: _type = "Task" [ 996.652931] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.660949] env[62385]: DEBUG oslo_vmware.api [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5234f616-8fab-3ae3-54c4-53469f3acb39, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.799433] env[62385]: DEBUG oslo_concurrency.lockutils [req-b5e7d32f-a810-45dd-8a6b-3be0698e4b48 req-83a6d099-a3b5-40e3-b3bd-4d58287edffa service nova] Releasing lock "refresh_cache-784589bb-a668-4674-83b8-6219def15067" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.836635] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9eff707c-16e0-46c5-b85d-791396fd0db5 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "5daf0a99-0c2a-40d8-afc7-1998e21e32d8" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.012s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.841408] env[62385]: DEBUG oslo_vmware.api [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206275, 'name': ReconfigVM_Task, 'duration_secs': 0.573606} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.842462] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Reconfigured VM instance instance-00000064 to attach disk [datastore1] 76672404-2bcf-438a-9e21-92ea9dc86461/76672404-2bcf-438a-9e21-92ea9dc86461.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 996.845349] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-80a9072d-13d3-48b8-8cd5-3556dc57901c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.854833] env[62385]: DEBUG oslo_vmware.api [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 996.854833] env[62385]: value = "task-1206277" [ 996.854833] env[62385]: _type = "Task" [ 996.854833] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.864613] env[62385]: DEBUG oslo_vmware.api [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206277, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.902177] env[62385]: DEBUG nova.network.neutron [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 996.979666] env[62385]: DEBUG nova.network.neutron [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Updating instance_info_cache with network_info: [{"id": "404c104c-258f-480e-bfe2-4ace4e83a43c", "address": "fa:16:3e:87:f8:2a", "network": {"id": "06017161-984c-4657-9639-4e6bdb6d4c09", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1072882119-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.208", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d45b5346eab44707b4d81eaaf12b0935", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39cd75b0-9ec7-48ed-b57f-34da0c573a60", "external-id": "nsx-vlan-transportzone-751", "segmentation_id": 751, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap404c104c-25", "ovs_interfaceid": "404c104c-258f-480e-bfe2-4ace4e83a43c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.066302] env[62385]: DEBUG nova.scheduler.client.report [None req-a3a6fc1a-c0ca-4fe5-9a14-65b9435d0e5f tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 997.099323] env[62385]: DEBUG oslo_concurrency.lockutils [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Releasing lock "refresh_cache-b1e0baf3-643c-49c2-8a80-8de07d8527e7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.134152] env[62385]: DEBUG nova.network.neutron [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Updating instance_info_cache with network_info: [{"id": "a7153a2d-db69-4d14-8ddc-b634ddabbc18", "address": "fa:16:3e:a3:0d:ab", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7153a2d-db", "ovs_interfaceid": "a7153a2d-db69-4d14-8ddc-b634ddabbc18", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.165558] env[62385]: DEBUG oslo_vmware.api [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5234f616-8fab-3ae3-54c4-53469f3acb39, 'name': SearchDatastore_Task, 'duration_secs': 0.008144} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.166323] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2bfc8dde-8d50-4f92-918e-04929397f505 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.172452] env[62385]: DEBUG oslo_vmware.api [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Waiting for the task: (returnval){ [ 997.172452] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]523af282-e934-2319-5991-404b80a587ff" [ 997.172452] env[62385]: _type = "Task" [ 997.172452] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.181472] env[62385]: DEBUG oslo_vmware.api [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]523af282-e934-2319-5991-404b80a587ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.371988] env[62385]: DEBUG oslo_vmware.api [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206277, 'name': Rename_Task, 'duration_secs': 0.149485} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.372832] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 997.373152] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1ff381e0-1dcf-416d-885f-4be2db784650 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.380721] env[62385]: DEBUG oslo_vmware.api [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 997.380721] env[62385]: value = "task-1206278" [ 997.380721] env[62385]: _type = "Task" [ 997.380721] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.389306] env[62385]: DEBUG oslo_vmware.api [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206278, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.445332] env[62385]: DEBUG nova.compute.manager [req-ab8377ff-febc-4c41-829d-1551c40f8b74 req-8303de40-feb9-46ee-9ebe-038b3745f785 service nova] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Received event network-vif-plugged-a7153a2d-db69-4d14-8ddc-b634ddabbc18 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 997.445494] env[62385]: DEBUG oslo_concurrency.lockutils [req-ab8377ff-febc-4c41-829d-1551c40f8b74 req-8303de40-feb9-46ee-9ebe-038b3745f785 service nova] Acquiring lock "57637d6e-39e2-405c-bbcf-64bf0ffbebd1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.445700] env[62385]: DEBUG oslo_concurrency.lockutils [req-ab8377ff-febc-4c41-829d-1551c40f8b74 req-8303de40-feb9-46ee-9ebe-038b3745f785 service nova] Lock "57637d6e-39e2-405c-bbcf-64bf0ffbebd1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.446046] env[62385]: DEBUG oslo_concurrency.lockutils [req-ab8377ff-febc-4c41-829d-1551c40f8b74 req-8303de40-feb9-46ee-9ebe-038b3745f785 service nova] Lock "57637d6e-39e2-405c-bbcf-64bf0ffbebd1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.446046] env[62385]: DEBUG nova.compute.manager [req-ab8377ff-febc-4c41-829d-1551c40f8b74 req-8303de40-feb9-46ee-9ebe-038b3745f785 service nova] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] No waiting events found dispatching network-vif-plugged-a7153a2d-db69-4d14-8ddc-b634ddabbc18 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 997.446233] env[62385]: WARNING nova.compute.manager [req-ab8377ff-febc-4c41-829d-1551c40f8b74 req-8303de40-feb9-46ee-9ebe-038b3745f785 service nova] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Received unexpected event network-vif-plugged-a7153a2d-db69-4d14-8ddc-b634ddabbc18 for instance with vm_state building and task_state spawning. [ 997.446370] env[62385]: DEBUG nova.compute.manager [req-ab8377ff-febc-4c41-829d-1551c40f8b74 req-8303de40-feb9-46ee-9ebe-038b3745f785 service nova] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Received event network-changed-a7153a2d-db69-4d14-8ddc-b634ddabbc18 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 997.446545] env[62385]: DEBUG nova.compute.manager [req-ab8377ff-febc-4c41-829d-1551c40f8b74 req-8303de40-feb9-46ee-9ebe-038b3745f785 service nova] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Refreshing instance network info cache due to event network-changed-a7153a2d-db69-4d14-8ddc-b634ddabbc18. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 997.446774] env[62385]: DEBUG oslo_concurrency.lockutils [req-ab8377ff-febc-4c41-829d-1551c40f8b74 req-8303de40-feb9-46ee-9ebe-038b3745f785 service nova] Acquiring lock "refresh_cache-57637d6e-39e2-405c-bbcf-64bf0ffbebd1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.484883] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Releasing lock "refresh_cache-9b9fd167-d843-4b9d-9e4f-dbde5d3d629e" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.574754] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a3a6fc1a-c0ca-4fe5-9a14-65b9435d0e5f tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.846s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.596948] env[62385]: INFO nova.scheduler.client.report [None req-a3a6fc1a-c0ca-4fe5-9a14-65b9435d0e5f tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Deleted allocations for instance 9741e9c8-60d9-41bf-8185-ef3cfcc1c844 [ 997.622480] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4247193a-63b4-44dd-b1a9-68c3cc827ecd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.642386] env[62385]: DEBUG oslo_concurrency.lockutils [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Releasing lock "refresh_cache-57637d6e-39e2-405c-bbcf-64bf0ffbebd1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.643091] env[62385]: DEBUG nova.compute.manager [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Instance network_info: |[{"id": "a7153a2d-db69-4d14-8ddc-b634ddabbc18", "address": "fa:16:3e:a3:0d:ab", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7153a2d-db", "ovs_interfaceid": "a7153a2d-db69-4d14-8ddc-b634ddabbc18", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 997.643295] env[62385]: DEBUG oslo_concurrency.lockutils [req-ab8377ff-febc-4c41-829d-1551c40f8b74 req-8303de40-feb9-46ee-9ebe-038b3745f785 service nova] Acquired lock "refresh_cache-57637d6e-39e2-405c-bbcf-64bf0ffbebd1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.643615] env[62385]: DEBUG nova.network.neutron [req-ab8377ff-febc-4c41-829d-1551c40f8b74 req-8303de40-feb9-46ee-9ebe-038b3745f785 service nova] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Refreshing network info cache for port a7153a2d-db69-4d14-8ddc-b634ddabbc18 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 997.644932] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a3:0d:ab', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '685b4083-b748-41fb-a68a-273b1073fa28', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a7153a2d-db69-4d14-8ddc-b634ddabbc18', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 997.652346] env[62385]: DEBUG oslo.service.loopingcall [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 997.653066] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-302166c3-d35a-4d26-a8bb-24ec952a52a1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.658735] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 997.659881] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8807de7d-c43d-479a-bfe6-c8c404264cc3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.682795] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Updating instance 'b1e0baf3-643c-49c2-8a80-8de07d8527e7' progress to 83 {{(pid=62385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 997.688200] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 997.688200] env[62385]: value = "task-1206279" [ 997.688200] env[62385]: _type = "Task" [ 997.688200] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.696423] env[62385]: DEBUG oslo_vmware.api [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]523af282-e934-2319-5991-404b80a587ff, 'name': SearchDatastore_Task, 'duration_secs': 0.009782} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.696670] env[62385]: DEBUG oslo_concurrency.lockutils [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.696959] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 784589bb-a668-4674-83b8-6219def15067/784589bb-a668-4674-83b8-6219def15067.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 997.697247] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-792be39b-4082-43d1-91c8-c2e0d0db2b83 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.702700] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206279, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.709188] env[62385]: DEBUG oslo_vmware.api [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Waiting for the task: (returnval){ [ 997.709188] env[62385]: value = "task-1206280" [ 997.709188] env[62385]: _type = "Task" [ 997.709188] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.860131] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9eff707c-16e0-46c5-b85d-791396fd0db5 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "5daf0a99-0c2a-40d8-afc7-1998e21e32d8" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.860333] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9eff707c-16e0-46c5-b85d-791396fd0db5 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "5daf0a99-0c2a-40d8-afc7-1998e21e32d8" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.860596] env[62385]: INFO nova.compute.manager [None req-9eff707c-16e0-46c5-b85d-791396fd0db5 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Attaching volume 9621f12c-980a-4762-8ea2-c4fff11dc20e to /dev/sdb [ 997.901546] env[62385]: DEBUG oslo_vmware.api [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206278, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.917156] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aca91df-caaa-42b1-9c47-2e52e757a85a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.924443] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94bfdc4c-e5ce-4372-9048-35c813716174 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.944039] env[62385]: DEBUG nova.virt.block_device [None req-9eff707c-16e0-46c5-b85d-791396fd0db5 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Updating existing volume attachment record: 8926b1a5-1212-45e9-a4f7-805bbc25f885 {{(pid=62385) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 998.074747] env[62385]: DEBUG nova.network.neutron [None req-8e10c3aa-33db-4d4b-ae34-41f4b8a0ebec tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Updating instance_info_cache with network_info: [{"id": "a9513285-4023-4f3e-b90d-3eb507955f13", "address": "fa:16:3e:30:02:c0", "network": {"id": "30c4e78b-6a64-47dc-aaaa-a210160c9f23", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1809587656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "105150c170d74e268e82ab894e9bf0d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10ff2092-e8eb-4768-ad4a-65a80560b447", "external-id": "nsx-vlan-transportzone-845", "segmentation_id": 845, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa9513285-40", "ovs_interfaceid": "a9513285-4023-4f3e-b90d-3eb507955f13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.108151] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a3a6fc1a-c0ca-4fe5-9a14-65b9435d0e5f tempest-ImagesTestJSON-1011004794 tempest-ImagesTestJSON-1011004794-project-member] Lock "9741e9c8-60d9-41bf-8185-ef3cfcc1c844" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.229s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.192482] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 998.192482] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2fa46772-6ff4-41e3-b554-d0a78f98c621 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.214256] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206279, 'name': CreateVM_Task} progress is 25%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.223680] env[62385]: DEBUG oslo_vmware.api [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 998.223680] env[62385]: value = "task-1206283" [ 998.223680] env[62385]: _type = "Task" [ 998.223680] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.233348] env[62385]: DEBUG oslo_vmware.api [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Task: {'id': task-1206280, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.242158] env[62385]: DEBUG oslo_vmware.api [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206283, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.396690] env[62385]: DEBUG oslo_vmware.api [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206278, 'name': PowerOnVM_Task, 'duration_secs': 0.568417} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.396690] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 998.396690] env[62385]: INFO nova.compute.manager [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Took 7.71 seconds to spawn the instance on the hypervisor. [ 998.396690] env[62385]: DEBUG nova.compute.manager [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 998.396690] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17644c82-87d0-4c2e-9560-02e3cf17409b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.533125] env[62385]: DEBUG nova.network.neutron [req-ab8377ff-febc-4c41-829d-1551c40f8b74 req-8303de40-feb9-46ee-9ebe-038b3745f785 service nova] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Updated VIF entry in instance network info cache for port a7153a2d-db69-4d14-8ddc-b634ddabbc18. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 998.533535] env[62385]: DEBUG nova.network.neutron [req-ab8377ff-febc-4c41-829d-1551c40f8b74 req-8303de40-feb9-46ee-9ebe-038b3745f785 service nova] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Updating instance_info_cache with network_info: [{"id": "a7153a2d-db69-4d14-8ddc-b634ddabbc18", "address": "fa:16:3e:a3:0d:ab", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7153a2d-db", "ovs_interfaceid": "a7153a2d-db69-4d14-8ddc-b634ddabbc18", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.580811] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8e10c3aa-33db-4d4b-ae34-41f4b8a0ebec tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Releasing lock "refresh_cache-df020d6b-3fab-4599-a342-47c7833b4240" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.581245] env[62385]: DEBUG nova.objects.instance [None req-8e10c3aa-33db-4d4b-ae34-41f4b8a0ebec tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lazy-loading 'migration_context' on Instance uuid df020d6b-3fab-4599-a342-47c7833b4240 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 998.703424] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206279, 'name': CreateVM_Task, 'duration_secs': 0.697697} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.703587] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 998.704324] env[62385]: DEBUG oslo_concurrency.lockutils [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 998.704609] env[62385]: DEBUG oslo_concurrency.lockutils [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.704962] env[62385]: DEBUG oslo_concurrency.lockutils [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 998.705260] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dba08a7b-3abc-433c-a5d8-cb107ab2b74d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.720258] env[62385]: DEBUG oslo_vmware.api [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 998.720258] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52f839c0-684e-ce2d-8c31-d0ab36af1024" [ 998.720258] env[62385]: _type = "Task" [ 998.720258] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.727512] env[62385]: DEBUG oslo_vmware.api [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Task: {'id': task-1206280, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.541003} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.731231] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 784589bb-a668-4674-83b8-6219def15067/784589bb-a668-4674-83b8-6219def15067.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 998.731506] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] [instance: 784589bb-a668-4674-83b8-6219def15067] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 998.734651] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c63137b1-a2c9-4d65-bcde-f22c45a4288d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.736437] env[62385]: DEBUG oslo_vmware.api [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52f839c0-684e-ce2d-8c31-d0ab36af1024, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.742929] env[62385]: DEBUG oslo_vmware.api [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206283, 'name': PowerOnVM_Task, 'duration_secs': 0.461945} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.745143] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 998.745443] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-956cca75-a279-415c-a2bc-0b9668d30c57 tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Updating instance 'b1e0baf3-643c-49c2-8a80-8de07d8527e7' progress to 100 {{(pid=62385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 998.750112] env[62385]: DEBUG oslo_vmware.api [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Waiting for the task: (returnval){ [ 998.750112] env[62385]: value = "task-1206286" [ 998.750112] env[62385]: _type = "Task" [ 998.750112] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.760127] env[62385]: DEBUG oslo_vmware.api [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Task: {'id': task-1206286, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.915474] env[62385]: INFO nova.compute.manager [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Took 18.60 seconds to build instance. [ 998.999655] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ac68465-ca04-4097-bad7-c85b89804d51 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.019609] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Updating instance '9b9fd167-d843-4b9d-9e4f-dbde5d3d629e' progress to 0 {{(pid=62385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 999.036481] env[62385]: DEBUG oslo_concurrency.lockutils [req-ab8377ff-febc-4c41-829d-1551c40f8b74 req-8303de40-feb9-46ee-9ebe-038b3745f785 service nova] Releasing lock "refresh_cache-57637d6e-39e2-405c-bbcf-64bf0ffbebd1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.084706] env[62385]: DEBUG nova.objects.base [None req-8e10c3aa-33db-4d4b-ae34-41f4b8a0ebec tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62385) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 999.085711] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-561b7a4c-ca05-4766-9bf3-338a7986cf60 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.106324] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-552e5198-3cee-46f7-bc1e-ca8112c47130 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.114403] env[62385]: DEBUG oslo_vmware.api [None req-8e10c3aa-33db-4d4b-ae34-41f4b8a0ebec tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 999.114403] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5256499a-fe0c-20b9-e52e-33a4bedf790e" [ 999.114403] env[62385]: _type = "Task" [ 999.114403] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.120832] env[62385]: DEBUG oslo_vmware.api [None req-8e10c3aa-33db-4d4b-ae34-41f4b8a0ebec tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5256499a-fe0c-20b9-e52e-33a4bedf790e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.235125] env[62385]: DEBUG oslo_vmware.api [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52f839c0-684e-ce2d-8c31-d0ab36af1024, 'name': SearchDatastore_Task, 'duration_secs': 0.02396} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.235125] env[62385]: DEBUG oslo_concurrency.lockutils [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.235125] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 999.235125] env[62385]: DEBUG oslo_concurrency.lockutils [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.235125] env[62385]: DEBUG oslo_concurrency.lockutils [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.235520] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 999.236151] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b86e7bdd-28ca-47ff-804a-ea60650897f2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.248327] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 999.249142] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 999.251075] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4849c966-99e5-40df-ad3a-f548de40097f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.271859] env[62385]: DEBUG oslo_vmware.api [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 999.271859] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5287821b-d64c-0fe6-2b2e-1b146c1e136b" [ 999.271859] env[62385]: _type = "Task" [ 999.271859] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.275779] env[62385]: DEBUG oslo_vmware.api [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Task: {'id': task-1206286, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075543} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.280062] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] [instance: 784589bb-a668-4674-83b8-6219def15067] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 999.281139] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d46ec31a-f747-4acf-bafd-7a25e4a245b0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.312281] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] [instance: 784589bb-a668-4674-83b8-6219def15067] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] 784589bb-a668-4674-83b8-6219def15067/784589bb-a668-4674-83b8-6219def15067.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 999.316674] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-33019ba2-6a25-47b5-bfbd-3fcfcfbf3ca5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.332094] env[62385]: DEBUG oslo_vmware.api [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5287821b-d64c-0fe6-2b2e-1b146c1e136b, 'name': SearchDatastore_Task, 'duration_secs': 0.014538} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.333629] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aec6adbb-a482-49b3-bd55-8774c82675ed {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.340259] env[62385]: DEBUG oslo_vmware.api [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Waiting for the task: (returnval){ [ 999.340259] env[62385]: value = "task-1206287" [ 999.340259] env[62385]: _type = "Task" [ 999.340259] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.341936] env[62385]: DEBUG oslo_vmware.api [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 999.341936] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e88df7-4794-0d9b-aff5-7ccdcd095734" [ 999.341936] env[62385]: _type = "Task" [ 999.341936] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.353977] env[62385]: DEBUG oslo_vmware.api [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Task: {'id': task-1206287, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.357401] env[62385]: DEBUG oslo_vmware.api [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e88df7-4794-0d9b-aff5-7ccdcd095734, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.418120] env[62385]: DEBUG oslo_concurrency.lockutils [None req-95d33fa8-df11-432d-821b-f30708b2ddd2 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "76672404-2bcf-438a-9e21-92ea9dc86461" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.112s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.525605] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 999.525949] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-93ae0804-a808-436c-912f-b904126a6cde {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.536741] env[62385]: DEBUG oslo_vmware.api [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 999.536741] env[62385]: value = "task-1206288" [ 999.536741] env[62385]: _type = "Task" [ 999.536741] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.546848] env[62385]: DEBUG oslo_vmware.api [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206288, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.626441] env[62385]: DEBUG oslo_vmware.api [None req-8e10c3aa-33db-4d4b-ae34-41f4b8a0ebec tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5256499a-fe0c-20b9-e52e-33a4bedf790e, 'name': SearchDatastore_Task, 'duration_secs': 0.021258} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.626787] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8e10c3aa-33db-4d4b-ae34-41f4b8a0ebec tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.627446] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8e10c3aa-33db-4d4b-ae34-41f4b8a0ebec tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.859213] env[62385]: DEBUG oslo_vmware.api [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e88df7-4794-0d9b-aff5-7ccdcd095734, 'name': SearchDatastore_Task, 'duration_secs': 0.015079} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.862791] env[62385]: DEBUG oslo_concurrency.lockutils [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.863869] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 57637d6e-39e2-405c-bbcf-64bf0ffbebd1/57637d6e-39e2-405c-bbcf-64bf0ffbebd1.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 999.863869] env[62385]: DEBUG oslo_vmware.api [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Task: {'id': task-1206287, 'name': ReconfigVM_Task, 'duration_secs': 0.414717} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.863869] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ba3d83e3-c992-48a2-910a-07fc6566821f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.865933] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] [instance: 784589bb-a668-4674-83b8-6219def15067] Reconfigured VM instance instance-00000065 to attach disk [datastore1] 784589bb-a668-4674-83b8-6219def15067/784589bb-a668-4674-83b8-6219def15067.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 999.868048] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f096ae6d-2ff0-4941-85e5-9ba56998b31f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.879425] env[62385]: DEBUG oslo_vmware.api [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 999.879425] env[62385]: value = "task-1206290" [ 999.879425] env[62385]: _type = "Task" [ 999.879425] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.881127] env[62385]: DEBUG oslo_vmware.api [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Waiting for the task: (returnval){ [ 999.881127] env[62385]: value = "task-1206289" [ 999.881127] env[62385]: _type = "Task" [ 999.881127] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.900199] env[62385]: DEBUG oslo_vmware.api [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Task: {'id': task-1206289, 'name': Rename_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.903986] env[62385]: DEBUG oslo_vmware.api [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206290, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.053462] env[62385]: DEBUG oslo_vmware.api [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206288, 'name': PowerOffVM_Task, 'duration_secs': 0.289264} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.054566] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1000.054875] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Updating instance '9b9fd167-d843-4b9d-9e4f-dbde5d3d629e' progress to 17 {{(pid=62385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1000.397644] env[62385]: DEBUG oslo_vmware.api [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206290, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.409379] env[62385]: DEBUG oslo_vmware.api [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Task: {'id': task-1206289, 'name': Rename_Task, 'duration_secs': 0.209338} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.409756] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] [instance: 784589bb-a668-4674-83b8-6219def15067] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1000.410132] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ffe3c42a-7688-4804-b998-9286f72a0096 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.416807] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a51c3a05-fbd7-4283-9eb3-f7e1759db437 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.422674] env[62385]: DEBUG oslo_vmware.api [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Waiting for the task: (returnval){ [ 1000.422674] env[62385]: value = "task-1206293" [ 1000.422674] env[62385]: _type = "Task" [ 1000.422674] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.444855] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fdf00f8-a722-4777-b089-818f8dbb882c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.449417] env[62385]: DEBUG oslo_vmware.api [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Task: {'id': task-1206293, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.488297] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e1379ac-741e-42f5-972c-7e02ce014bed {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.498086] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fb9cbeb-43e1-40a4-80ef-61b9d9c491cc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.520233] env[62385]: DEBUG nova.compute.provider_tree [None req-8e10c3aa-33db-4d4b-ae34-41f4b8a0ebec tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1000.537359] env[62385]: DEBUG oslo_vmware.rw_handles [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52515341-9ab6-3170-21a5-eb2804418e3d/disk-0.vmdk. {{(pid=62385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1000.539007] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e5b0c56-a10d-444b-8c5b-1932191693b8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.545194] env[62385]: DEBUG oslo_concurrency.lockutils [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "76672404-2bcf-438a-9e21-92ea9dc86461" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.545590] env[62385]: DEBUG oslo_concurrency.lockutils [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "76672404-2bcf-438a-9e21-92ea9dc86461" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.545949] env[62385]: DEBUG oslo_concurrency.lockutils [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "76672404-2bcf-438a-9e21-92ea9dc86461-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.546269] env[62385]: DEBUG oslo_concurrency.lockutils [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "76672404-2bcf-438a-9e21-92ea9dc86461-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.546527] env[62385]: DEBUG oslo_concurrency.lockutils [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "76672404-2bcf-438a-9e21-92ea9dc86461-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.551138] env[62385]: DEBUG oslo_vmware.rw_handles [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52515341-9ab6-3170-21a5-eb2804418e3d/disk-0.vmdk is in state: ready. {{(pid=62385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1000.551379] env[62385]: ERROR oslo_vmware.rw_handles [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52515341-9ab6-3170-21a5-eb2804418e3d/disk-0.vmdk due to incomplete transfer. [ 1000.552341] env[62385]: INFO nova.compute.manager [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Terminating instance [ 1000.554283] env[62385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-51f864c5-81cf-4ce3-a622-2d28efa2a588 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.559434] env[62385]: DEBUG nova.compute.manager [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1000.559738] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1000.562613] env[62385]: DEBUG nova.virt.hardware [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:16:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1000.562933] env[62385]: DEBUG nova.virt.hardware [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1000.563195] env[62385]: DEBUG nova.virt.hardware [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1000.563484] env[62385]: DEBUG nova.virt.hardware [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1000.563727] env[62385]: DEBUG nova.virt.hardware [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1000.563956] env[62385]: DEBUG nova.virt.hardware [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1000.564315] env[62385]: DEBUG nova.virt.hardware [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1000.564597] env[62385]: DEBUG nova.virt.hardware [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1000.564907] env[62385]: DEBUG nova.virt.hardware [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1000.565197] env[62385]: DEBUG nova.virt.hardware [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1000.565511] env[62385]: DEBUG nova.virt.hardware [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1000.574618] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9225953f-6151-47d7-9542-ea0138e783d0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.579990] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e1ca8f3a-6fa1-4128-abeb-9679ec64ceb6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.600389] env[62385]: DEBUG oslo_vmware.rw_handles [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52515341-9ab6-3170-21a5-eb2804418e3d/disk-0.vmdk. {{(pid=62385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1000.600807] env[62385]: DEBUG nova.virt.vmwareapi.images [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Uploaded image a561779b-d9f7-4373-8a4d-586fbc25c6d7 to the Glance image server {{(pid=62385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1000.604572] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Destroying the VM {{(pid=62385) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1000.608912] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-b6b36152-f25d-4928-9b5f-7a05671ef2b9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.611630] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1000.612486] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fd2c185e-8fd2-4fdc-bdc6-50f3c06b27f5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.616851] env[62385]: DEBUG oslo_vmware.api [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1000.616851] env[62385]: value = "task-1206294" [ 1000.616851] env[62385]: _type = "Task" [ 1000.616851] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.634127] env[62385]: DEBUG oslo_vmware.api [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206294, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.634657] env[62385]: DEBUG oslo_vmware.api [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1000.634657] env[62385]: value = "task-1206295" [ 1000.634657] env[62385]: _type = "Task" [ 1000.634657] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.636882] env[62385]: DEBUG oslo_vmware.api [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 1000.636882] env[62385]: value = "task-1206296" [ 1000.636882] env[62385]: _type = "Task" [ 1000.636882] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.659343] env[62385]: DEBUG oslo_vmware.api [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206296, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.659749] env[62385]: DEBUG oslo_vmware.api [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206295, 'name': Destroy_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.893892] env[62385]: DEBUG oslo_vmware.api [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206290, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.612645} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.894241] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 57637d6e-39e2-405c-bbcf-64bf0ffbebd1/57637d6e-39e2-405c-bbcf-64bf0ffbebd1.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1000.894477] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1000.894777] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b542fd0d-5ec3-4eb2-a2cf-856bd4d25a20 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.903023] env[62385]: DEBUG oslo_vmware.api [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 1000.903023] env[62385]: value = "task-1206297" [ 1000.903023] env[62385]: _type = "Task" [ 1000.903023] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.911775] env[62385]: DEBUG oslo_vmware.api [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206297, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.938913] env[62385]: DEBUG oslo_vmware.api [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Task: {'id': task-1206293, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.024127] env[62385]: DEBUG nova.scheduler.client.report [None req-8e10c3aa-33db-4d4b-ae34-41f4b8a0ebec tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1001.129611] env[62385]: DEBUG oslo_vmware.api [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206294, 'name': ReconfigVM_Task, 'duration_secs': 0.21272} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.129915] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Updating instance '9b9fd167-d843-4b9d-9e4f-dbde5d3d629e' progress to 33 {{(pid=62385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1001.145351] env[62385]: DEBUG oslo_vmware.api [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206295, 'name': Destroy_Task} progress is 33%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.153734] env[62385]: DEBUG oslo_vmware.api [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206296, 'name': PowerOffVM_Task, 'duration_secs': 0.218361} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.154021] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1001.154208] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1001.154466] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-289a7753-3782-44cd-aaa8-754ccd9e93be {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.231666] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1001.231908] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1001.232209] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Deleting the datastore file [datastore1] 76672404-2bcf-438a-9e21-92ea9dc86461 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1001.232513] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7f19ebca-61ea-4c79-86d8-4ec2d2cff406 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.240704] env[62385]: DEBUG oslo_vmware.api [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 1001.240704] env[62385]: value = "task-1206299" [ 1001.240704] env[62385]: _type = "Task" [ 1001.240704] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.249326] env[62385]: DEBUG oslo_vmware.api [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206299, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.414019] env[62385]: DEBUG oslo_vmware.api [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206297, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.120946} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.414310] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1001.415103] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1303f7f5-d8d7-4056-87e6-b63fca5010f6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.440021] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] 57637d6e-39e2-405c-bbcf-64bf0ffbebd1/57637d6e-39e2-405c-bbcf-64bf0ffbebd1.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1001.443753] env[62385]: DEBUG oslo_concurrency.lockutils [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "b1e0baf3-643c-49c2-8a80-8de07d8527e7" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.443982] env[62385]: DEBUG oslo_concurrency.lockutils [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "b1e0baf3-643c-49c2-8a80-8de07d8527e7" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.444182] env[62385]: DEBUG nova.compute.manager [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Going to confirm migration 3 {{(pid=62385) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1001.445625] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4af28929-be53-493a-8e70-d7367075f991 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.469127] env[62385]: DEBUG oslo_vmware.api [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Task: {'id': task-1206293, 'name': PowerOnVM_Task, 'duration_secs': 0.612469} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.470752] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] [instance: 784589bb-a668-4674-83b8-6219def15067] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1001.470752] env[62385]: INFO nova.compute.manager [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] [instance: 784589bb-a668-4674-83b8-6219def15067] Took 8.40 seconds to spawn the instance on the hypervisor. [ 1001.470920] env[62385]: DEBUG nova.compute.manager [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] [instance: 784589bb-a668-4674-83b8-6219def15067] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1001.471243] env[62385]: DEBUG oslo_vmware.api [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 1001.471243] env[62385]: value = "task-1206300" [ 1001.471243] env[62385]: _type = "Task" [ 1001.471243] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.471957] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d2bf97f-8ef3-4b4d-bb79-92ffb1f29442 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.487255] env[62385]: DEBUG oslo_vmware.api [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206300, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.636909] env[62385]: DEBUG nova.virt.hardware [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1001.637127] env[62385]: DEBUG nova.virt.hardware [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1001.637329] env[62385]: DEBUG nova.virt.hardware [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1001.637620] env[62385]: DEBUG nova.virt.hardware [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1001.637972] env[62385]: DEBUG nova.virt.hardware [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1001.638254] env[62385]: DEBUG nova.virt.hardware [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1001.638522] env[62385]: DEBUG nova.virt.hardware [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1001.638702] env[62385]: DEBUG nova.virt.hardware [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1001.638886] env[62385]: DEBUG nova.virt.hardware [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1001.639078] env[62385]: DEBUG nova.virt.hardware [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1001.639269] env[62385]: DEBUG nova.virt.hardware [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1001.645169] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Reconfiguring VM instance instance-00000044 to detach disk 2000 {{(pid=62385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1001.645483] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fa64d15c-8bbe-412b-8800-4515d1926290 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.667148] env[62385]: DEBUG oslo_vmware.api [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206295, 'name': Destroy_Task, 'duration_secs': 0.665594} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.668467] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Destroyed the VM [ 1001.668718] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Deleting Snapshot of the VM instance {{(pid=62385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1001.669463] env[62385]: DEBUG oslo_vmware.api [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1001.669463] env[62385]: value = "task-1206301" [ 1001.669463] env[62385]: _type = "Task" [ 1001.669463] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.669463] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-77ceccbd-dd1a-4348-9bca-a85ab14529e3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.679978] env[62385]: DEBUG oslo_vmware.api [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206301, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.681277] env[62385]: DEBUG oslo_vmware.api [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1001.681277] env[62385]: value = "task-1206302" [ 1001.681277] env[62385]: _type = "Task" [ 1001.681277] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.690345] env[62385]: DEBUG oslo_vmware.api [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206302, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.751515] env[62385]: DEBUG oslo_vmware.api [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206299, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.502624} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.751801] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1001.752023] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1001.752219] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1001.752401] env[62385]: INFO nova.compute.manager [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Took 1.19 seconds to destroy the instance on the hypervisor. [ 1001.752646] env[62385]: DEBUG oslo.service.loopingcall [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1001.752848] env[62385]: DEBUG nova.compute.manager [-] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1001.752995] env[62385]: DEBUG nova.network.neutron [-] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1001.986390] env[62385]: DEBUG oslo_vmware.api [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206300, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.002401] env[62385]: INFO nova.compute.manager [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] [instance: 784589bb-a668-4674-83b8-6219def15067] Took 17.88 seconds to build instance. [ 1002.033824] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8e10c3aa-33db-4d4b-ae34-41f4b8a0ebec tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.406s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.038471] env[62385]: DEBUG nova.compute.manager [req-6e065164-f45f-4d9a-a28a-cd811f59e6b6 req-ddef57c2-41ad-4c2b-9f05-4a8527b0aa96 service nova] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Received event network-vif-deleted-117fbbdf-9bdd-4d68-ac9f-d2fa6279aebd {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1002.039606] env[62385]: INFO nova.compute.manager [req-6e065164-f45f-4d9a-a28a-cd811f59e6b6 req-ddef57c2-41ad-4c2b-9f05-4a8527b0aa96 service nova] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Neutron deleted interface 117fbbdf-9bdd-4d68-ac9f-d2fa6279aebd; detaching it from the instance and deleting it from the info cache [ 1002.039677] env[62385]: DEBUG nova.network.neutron [req-6e065164-f45f-4d9a-a28a-cd811f59e6b6 req-ddef57c2-41ad-4c2b-9f05-4a8527b0aa96 service nova] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.042834] env[62385]: DEBUG oslo_concurrency.lockutils [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "refresh_cache-b1e0baf3-643c-49c2-8a80-8de07d8527e7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.043132] env[62385]: DEBUG oslo_concurrency.lockutils [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquired lock "refresh_cache-b1e0baf3-643c-49c2-8a80-8de07d8527e7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.043338] env[62385]: DEBUG nova.network.neutron [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1002.043516] env[62385]: DEBUG nova.objects.instance [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lazy-loading 'info_cache' on Instance uuid b1e0baf3-643c-49c2-8a80-8de07d8527e7 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1002.181433] env[62385]: DEBUG oslo_vmware.api [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206301, 'name': ReconfigVM_Task, 'duration_secs': 0.173827} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.181831] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Reconfigured VM instance instance-00000044 to detach disk 2000 {{(pid=62385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1002.185416] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0da22f6f-4105-4467-a035-508c425ebe06 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.193738] env[62385]: DEBUG oslo_vmware.api [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206302, 'name': RemoveSnapshot_Task, 'duration_secs': 0.400138} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.204333] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Deleted Snapshot of the VM instance {{(pid=62385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1002.204638] env[62385]: DEBUG nova.compute.manager [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1002.212281] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e/9b9fd167-d843-4b9d-9e4f-dbde5d3d629e.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1002.213049] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7c0019b-1beb-4203-84ed-42cfb651686a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.215690] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-88512dc6-02b9-4585-8a5d-02a51bfe0b28 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.238040] env[62385]: DEBUG oslo_vmware.api [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1002.238040] env[62385]: value = "task-1206304" [ 1002.238040] env[62385]: _type = "Task" [ 1002.238040] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.246729] env[62385]: DEBUG oslo_vmware.api [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206304, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.489886] env[62385]: DEBUG oslo_vmware.api [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206300, 'name': ReconfigVM_Task, 'duration_secs': 0.704449} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.490308] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Reconfigured VM instance instance-00000066 to attach disk [datastore1] 57637d6e-39e2-405c-bbcf-64bf0ffbebd1/57637d6e-39e2-405c-bbcf-64bf0ffbebd1.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1002.490889] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3dc905cc-b885-4081-bbb8-efd8a0cfc002 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.495960] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-9eff707c-16e0-46c5-b85d-791396fd0db5 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Volume attach. Driver type: vmdk {{(pid=62385) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1002.496542] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-9eff707c-16e0-46c5-b85d-791396fd0db5 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261287', 'volume_id': '9621f12c-980a-4762-8ea2-c4fff11dc20e', 'name': 'volume-9621f12c-980a-4762-8ea2-c4fff11dc20e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5daf0a99-0c2a-40d8-afc7-1998e21e32d8', 'attached_at': '', 'detached_at': '', 'volume_id': '9621f12c-980a-4762-8ea2-c4fff11dc20e', 'serial': '9621f12c-980a-4762-8ea2-c4fff11dc20e'} {{(pid=62385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1002.497419] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbd029ca-3d62-47f3-b7ad-4a438bde4525 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.501757] env[62385]: DEBUG oslo_vmware.api [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 1002.501757] env[62385]: value = "task-1206305" [ 1002.501757] env[62385]: _type = "Task" [ 1002.501757] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.506020] env[62385]: DEBUG oslo_concurrency.lockutils [None req-df943b43-d6a5-472c-b50d-af2f5fa0ee3e tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Lock "784589bb-a668-4674-83b8-6219def15067" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.390s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.506697] env[62385]: DEBUG nova.network.neutron [-] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.525711] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-695d8115-c862-46bb-ae63-f1313a945fbb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.528861] env[62385]: DEBUG oslo_vmware.api [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206305, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.561390] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-9eff707c-16e0-46c5-b85d-791396fd0db5 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] volume-9621f12c-980a-4762-8ea2-c4fff11dc20e/volume-9621f12c-980a-4762-8ea2-c4fff11dc20e.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1002.564686] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-71fbf0a0-1cda-459f-8544-6399b1aa3ec4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.567236] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-552ec5a4-2ea4-40de-815d-1f09c11956b9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.595600] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ce35763-fae7-41fe-bf81-f7c5999c61f0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.612740] env[62385]: DEBUG oslo_vmware.api [None req-9eff707c-16e0-46c5-b85d-791396fd0db5 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1002.612740] env[62385]: value = "task-1206306" [ 1002.612740] env[62385]: _type = "Task" [ 1002.612740] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.625267] env[62385]: DEBUG oslo_vmware.api [None req-9eff707c-16e0-46c5-b85d-791396fd0db5 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206306, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.639761] env[62385]: DEBUG nova.compute.manager [req-6e065164-f45f-4d9a-a28a-cd811f59e6b6 req-ddef57c2-41ad-4c2b-9f05-4a8527b0aa96 service nova] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Detach interface failed, port_id=117fbbdf-9bdd-4d68-ac9f-d2fa6279aebd, reason: Instance 76672404-2bcf-438a-9e21-92ea9dc86461 could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1002.652091] env[62385]: INFO nova.scheduler.client.report [None req-8e10c3aa-33db-4d4b-ae34-41f4b8a0ebec tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Deleted allocation for migration 183b505f-acf8-4ca5-bd5b-bbc88a8c7fd6 [ 1002.745662] env[62385]: INFO nova.compute.manager [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Shelve offloading [ 1002.747669] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1002.747813] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-012fde50-277c-445f-b3cb-5bd5896458cb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.755435] env[62385]: DEBUG oslo_vmware.api [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206304, 'name': ReconfigVM_Task, 'duration_secs': 0.425961} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.756740] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Reconfigured VM instance instance-00000044 to attach disk [datastore1] 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e/9b9fd167-d843-4b9d-9e4f-dbde5d3d629e.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1002.756984] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Updating instance '9b9fd167-d843-4b9d-9e4f-dbde5d3d629e' progress to 50 {{(pid=62385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1002.761105] env[62385]: DEBUG oslo_vmware.api [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1002.761105] env[62385]: value = "task-1206307" [ 1002.761105] env[62385]: _type = "Task" [ 1002.761105] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.771248] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] VM already powered off {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1002.771499] env[62385]: DEBUG nova.compute.manager [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1002.772369] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28b2899f-ae93-4e27-ac5f-07deb07f8198 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.779208] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquiring lock "refresh_cache-8e1d416c-2624-468c-94d7-6265e9f4178f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.779398] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquired lock "refresh_cache-8e1d416c-2624-468c-94d7-6265e9f4178f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.779594] env[62385]: DEBUG nova.network.neutron [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1003.013049] env[62385]: INFO nova.compute.manager [-] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Took 1.26 seconds to deallocate network for instance. [ 1003.013403] env[62385]: DEBUG oslo_vmware.api [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206305, 'name': Rename_Task, 'duration_secs': 0.223698} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.015146] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1003.015396] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6c13bb8d-6408-4a81-88bf-3ca024c6992d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.025891] env[62385]: DEBUG oslo_vmware.api [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 1003.025891] env[62385]: value = "task-1206308" [ 1003.025891] env[62385]: _type = "Task" [ 1003.025891] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.034104] env[62385]: DEBUG oslo_vmware.api [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206308, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.124490] env[62385]: DEBUG oslo_vmware.api [None req-9eff707c-16e0-46c5-b85d-791396fd0db5 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206306, 'name': ReconfigVM_Task, 'duration_secs': 0.415663} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.124845] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-9eff707c-16e0-46c5-b85d-791396fd0db5 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Reconfigured VM instance instance-0000005d to attach disk [datastore1] volume-9621f12c-980a-4762-8ea2-c4fff11dc20e/volume-9621f12c-980a-4762-8ea2-c4fff11dc20e.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1003.131800] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0519c676-f7c4-4237-8437-ce032bf7ccf7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.148034] env[62385]: DEBUG oslo_vmware.api [None req-9eff707c-16e0-46c5-b85d-791396fd0db5 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1003.148034] env[62385]: value = "task-1206309" [ 1003.148034] env[62385]: _type = "Task" [ 1003.148034] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.156527] env[62385]: DEBUG oslo_vmware.api [None req-9eff707c-16e0-46c5-b85d-791396fd0db5 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206309, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.159496] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8e10c3aa-33db-4d4b-ae34-41f4b8a0ebec tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "df020d6b-3fab-4599-a342-47c7833b4240" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.127s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.263828] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8aaca2c-558e-44ce-aaf2-32fa6c265cc3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.288070] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5275ac4a-a1d5-4c18-becf-3bf48473c8c2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.291439] env[62385]: DEBUG nova.network.neutron [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Updating instance_info_cache with network_info: [{"id": "eef524f4-fc00-47f1-a485-dcf568c6d0e4", "address": "fa:16:3e:e6:2c:55", "network": {"id": "ff21dd88-fa02-4bd6-8009-980ec6e72758", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1065789169-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aea48968d4464ac0ab01c998c059107d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3ff3baee-99ce-4b51-ae98-efc6163aaab3", "external-id": "nsx-vlan-transportzone-574", "segmentation_id": 574, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeef524f4-fc", "ovs_interfaceid": "eef524f4-fc00-47f1-a485-dcf568c6d0e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.309749] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Updating instance '9b9fd167-d843-4b9d-9e4f-dbde5d3d629e' progress to 67 {{(pid=62385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1003.522101] env[62385]: DEBUG oslo_concurrency.lockutils [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.522479] env[62385]: DEBUG oslo_concurrency.lockutils [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.523078] env[62385]: DEBUG nova.objects.instance [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lazy-loading 'resources' on Instance uuid 76672404-2bcf-438a-9e21-92ea9dc86461 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1003.540992] env[62385]: DEBUG oslo_vmware.api [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206308, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.563597] env[62385]: DEBUG nova.network.neutron [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Updating instance_info_cache with network_info: [{"id": "deafdeb9-f59c-48f4-873b-6e4b27d4a688", "address": "fa:16:3e:81:4d:75", "network": {"id": "9c4d6464-cacc-44f8-b7e5-0699cd312c53", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1889153615-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.186", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09c2c5e938bf46fd9353e47aecc72478", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4d3f69a-b086-4c3b-b976-5a848b63dfc4", "external-id": "nsx-vlan-transportzone-627", "segmentation_id": 627, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdeafdeb9-f5", "ovs_interfaceid": "deafdeb9-f59c-48f4-873b-6e4b27d4a688", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.658767] env[62385]: DEBUG oslo_vmware.api [None req-9eff707c-16e0-46c5-b85d-791396fd0db5 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206309, 'name': ReconfigVM_Task, 'duration_secs': 0.148277} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.659119] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-9eff707c-16e0-46c5-b85d-791396fd0db5 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261287', 'volume_id': '9621f12c-980a-4762-8ea2-c4fff11dc20e', 'name': 'volume-9621f12c-980a-4762-8ea2-c4fff11dc20e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5daf0a99-0c2a-40d8-afc7-1998e21e32d8', 'attached_at': '', 'detached_at': '', 'volume_id': '9621f12c-980a-4762-8ea2-c4fff11dc20e', 'serial': '9621f12c-980a-4762-8ea2-c4fff11dc20e'} {{(pid=62385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1003.794321] env[62385]: DEBUG oslo_concurrency.lockutils [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Releasing lock "refresh_cache-b1e0baf3-643c-49c2-8a80-8de07d8527e7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.794625] env[62385]: DEBUG nova.objects.instance [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lazy-loading 'migration_context' on Instance uuid b1e0baf3-643c-49c2-8a80-8de07d8527e7 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1003.864277] env[62385]: DEBUG nova.network.neutron [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Port 404c104c-258f-480e-bfe2-4ace4e83a43c binding to destination host cpu-1 is already ACTIVE {{(pid=62385) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1004.037900] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4becbe51-f49b-4c9e-99e8-a1178b22c7f1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "df020d6b-3fab-4599-a342-47c7833b4240" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.038168] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4becbe51-f49b-4c9e-99e8-a1178b22c7f1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "df020d6b-3fab-4599-a342-47c7833b4240" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.038380] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4becbe51-f49b-4c9e-99e8-a1178b22c7f1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "df020d6b-3fab-4599-a342-47c7833b4240-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.038565] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4becbe51-f49b-4c9e-99e8-a1178b22c7f1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "df020d6b-3fab-4599-a342-47c7833b4240-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.038737] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4becbe51-f49b-4c9e-99e8-a1178b22c7f1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "df020d6b-3fab-4599-a342-47c7833b4240-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.040486] env[62385]: DEBUG oslo_vmware.api [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206308, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.041082] env[62385]: INFO nova.compute.manager [None req-4becbe51-f49b-4c9e-99e8-a1178b22c7f1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Terminating instance [ 1004.043450] env[62385]: DEBUG nova.compute.manager [None req-4becbe51-f49b-4c9e-99e8-a1178b22c7f1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1004.043880] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4becbe51-f49b-4c9e-99e8-a1178b22c7f1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1004.046983] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d430c7ad-17c0-4a5b-93a1-bd50515375c0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.054976] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4becbe51-f49b-4c9e-99e8-a1178b22c7f1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1004.055235] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-59537bdf-8145-4af6-8445-caea65f9616f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.062098] env[62385]: DEBUG oslo_vmware.api [None req-4becbe51-f49b-4c9e-99e8-a1178b22c7f1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 1004.062098] env[62385]: value = "task-1206310" [ 1004.062098] env[62385]: _type = "Task" [ 1004.062098] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.065646] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Releasing lock "refresh_cache-8e1d416c-2624-468c-94d7-6265e9f4178f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.070902] env[62385]: DEBUG oslo_vmware.api [None req-4becbe51-f49b-4c9e-99e8-a1178b22c7f1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206310, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.224117] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2adabd35-59f0-4f26-aa20-55d3a1bb9f8d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.231215] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2833342c-42cb-4ae5-8a92-1364400e7267 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.263638] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a25afafc-a8ea-48e9-bbb4-079137152bf8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.271420] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92c5e723-1ea8-4aba-9fd2-ed53a6c59c04 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.286449] env[62385]: DEBUG nova.compute.provider_tree [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Updating inventory in ProviderTree for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1004.297074] env[62385]: DEBUG nova.objects.base [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62385) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1004.297990] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a09e26d6-6419-486f-a4ea-9f1daac732fc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.318442] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef8a5371-b725-4c51-8ba0-caf0728ca9db {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.324496] env[62385]: DEBUG oslo_vmware.api [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 1004.324496] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]525f94d8-33c5-ad14-d21a-719691c94ea4" [ 1004.324496] env[62385]: _type = "Task" [ 1004.324496] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.333944] env[62385]: DEBUG oslo_vmware.api [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]525f94d8-33c5-ad14-d21a-719691c94ea4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.456029] env[62385]: DEBUG nova.compute.manager [req-fa4d4abd-e648-4d7d-b888-44fee5183ae3 req-7f27e0db-4a8f-4664-b460-6eec80a8a3a7 service nova] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Received event network-vif-unplugged-deafdeb9-f59c-48f4-873b-6e4b27d4a688 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1004.456261] env[62385]: DEBUG oslo_concurrency.lockutils [req-fa4d4abd-e648-4d7d-b888-44fee5183ae3 req-7f27e0db-4a8f-4664-b460-6eec80a8a3a7 service nova] Acquiring lock "8e1d416c-2624-468c-94d7-6265e9f4178f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.456479] env[62385]: DEBUG oslo_concurrency.lockutils [req-fa4d4abd-e648-4d7d-b888-44fee5183ae3 req-7f27e0db-4a8f-4664-b460-6eec80a8a3a7 service nova] Lock "8e1d416c-2624-468c-94d7-6265e9f4178f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.456692] env[62385]: DEBUG oslo_concurrency.lockutils [req-fa4d4abd-e648-4d7d-b888-44fee5183ae3 req-7f27e0db-4a8f-4664-b460-6eec80a8a3a7 service nova] Lock "8e1d416c-2624-468c-94d7-6265e9f4178f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.456923] env[62385]: DEBUG nova.compute.manager [req-fa4d4abd-e648-4d7d-b888-44fee5183ae3 req-7f27e0db-4a8f-4664-b460-6eec80a8a3a7 service nova] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] No waiting events found dispatching network-vif-unplugged-deafdeb9-f59c-48f4-873b-6e4b27d4a688 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1004.457063] env[62385]: WARNING nova.compute.manager [req-fa4d4abd-e648-4d7d-b888-44fee5183ae3 req-7f27e0db-4a8f-4664-b460-6eec80a8a3a7 service nova] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Received unexpected event network-vif-unplugged-deafdeb9-f59c-48f4-873b-6e4b27d4a688 for instance with vm_state shelved and task_state shelving_offloading. [ 1004.511905] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1004.512987] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecc6304d-21c8-4bd7-b3ce-0546d23b2d85 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.523050] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1004.523470] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8316fe15-771b-42fe-b7f8-f4c2fdd8e1fa {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.539356] env[62385]: DEBUG oslo_vmware.api [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206308, 'name': PowerOnVM_Task, 'duration_secs': 1.406237} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.539356] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1004.539356] env[62385]: INFO nova.compute.manager [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Took 9.09 seconds to spawn the instance on the hypervisor. [ 1004.539898] env[62385]: DEBUG nova.compute.manager [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1004.541669] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-289579f5-ec04-41e6-bac5-480d92b91b51 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.547690] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "b0a6a3b9-05e4-4190-9cc6-4237dee07f91" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.547936] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "b0a6a3b9-05e4-4190-9cc6-4237dee07f91" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.573496] env[62385]: DEBUG oslo_vmware.api [None req-4becbe51-f49b-4c9e-99e8-a1178b22c7f1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206310, 'name': PowerOffVM_Task, 'duration_secs': 0.160666} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.573843] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4becbe51-f49b-4c9e-99e8-a1178b22c7f1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1004.574052] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4becbe51-f49b-4c9e-99e8-a1178b22c7f1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1004.574341] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3221e787-65aa-4433-8b1d-ce47708d6e52 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.596335] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1004.596666] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Deleting contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1004.596835] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Deleting the datastore file [datastore2] 8e1d416c-2624-468c-94d7-6265e9f4178f {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1004.597752] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f273188b-5562-4752-8ef4-9d1f70035245 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.605928] env[62385]: DEBUG oslo_vmware.api [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1004.605928] env[62385]: value = "task-1206313" [ 1004.605928] env[62385]: _type = "Task" [ 1004.605928] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.614690] env[62385]: DEBUG oslo_vmware.api [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206313, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.652067] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4becbe51-f49b-4c9e-99e8-a1178b22c7f1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1004.652067] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4becbe51-f49b-4c9e-99e8-a1178b22c7f1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Deleting contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1004.652291] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-4becbe51-f49b-4c9e-99e8-a1178b22c7f1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Deleting the datastore file [datastore2] df020d6b-3fab-4599-a342-47c7833b4240 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1004.652606] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fe9427c9-ea23-476a-8892-b15a0cbf6af5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.660047] env[62385]: DEBUG oslo_vmware.api [None req-4becbe51-f49b-4c9e-99e8-a1178b22c7f1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 1004.660047] env[62385]: value = "task-1206314" [ 1004.660047] env[62385]: _type = "Task" [ 1004.660047] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.669263] env[62385]: DEBUG oslo_vmware.api [None req-4becbe51-f49b-4c9e-99e8-a1178b22c7f1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206314, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.715881] env[62385]: DEBUG nova.objects.instance [None req-9eff707c-16e0-46c5-b85d-791396fd0db5 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lazy-loading 'flavor' on Instance uuid 5daf0a99-0c2a-40d8-afc7-1998e21e32d8 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1004.807236] env[62385]: ERROR nova.scheduler.client.report [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [req-84a4bea4-5099-45a3-9073-3c4e758ff3f0] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-84a4bea4-5099-45a3-9073-3c4e758ff3f0"}]} [ 1004.825312] env[62385]: DEBUG nova.scheduler.client.report [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Refreshing inventories for resource provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1004.836094] env[62385]: DEBUG oslo_vmware.api [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]525f94d8-33c5-ad14-d21a-719691c94ea4, 'name': SearchDatastore_Task, 'duration_secs': 0.008389} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.836384] env[62385]: DEBUG oslo_concurrency.lockutils [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.843528] env[62385]: DEBUG nova.scheduler.client.report [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Updating ProviderTree inventory for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1004.843744] env[62385]: DEBUG nova.compute.provider_tree [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Updating inventory in ProviderTree for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1004.855549] env[62385]: DEBUG nova.scheduler.client.report [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Refreshing aggregate associations for resource provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a, aggregates: None {{(pid=62385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1004.880713] env[62385]: DEBUG nova.scheduler.client.report [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Refreshing trait associations for resource provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=62385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1004.889084] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "9b9fd167-d843-4b9d-9e4f-dbde5d3d629e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.889319] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "9b9fd167-d843-4b9d-9e4f-dbde5d3d629e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.889498] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "9b9fd167-d843-4b9d-9e4f-dbde5d3d629e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.928170] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c637a0de-3a26-483f-8836-57b0fe6d6848 tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Acquiring lock "784589bb-a668-4674-83b8-6219def15067" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.928429] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c637a0de-3a26-483f-8836-57b0fe6d6848 tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Lock "784589bb-a668-4674-83b8-6219def15067" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.928639] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c637a0de-3a26-483f-8836-57b0fe6d6848 tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Acquiring lock "784589bb-a668-4674-83b8-6219def15067-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.928827] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c637a0de-3a26-483f-8836-57b0fe6d6848 tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Lock "784589bb-a668-4674-83b8-6219def15067-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.929040] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c637a0de-3a26-483f-8836-57b0fe6d6848 tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Lock "784589bb-a668-4674-83b8-6219def15067-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.931148] env[62385]: INFO nova.compute.manager [None req-c637a0de-3a26-483f-8836-57b0fe6d6848 tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] [instance: 784589bb-a668-4674-83b8-6219def15067] Terminating instance [ 1004.932975] env[62385]: DEBUG nova.compute.manager [None req-c637a0de-3a26-483f-8836-57b0fe6d6848 tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] [instance: 784589bb-a668-4674-83b8-6219def15067] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1004.933195] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c637a0de-3a26-483f-8836-57b0fe6d6848 tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] [instance: 784589bb-a668-4674-83b8-6219def15067] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1004.934041] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a50a1ea-82a2-4d5f-abb1-0c1ae3471638 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.946398] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c637a0de-3a26-483f-8836-57b0fe6d6848 tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] [instance: 784589bb-a668-4674-83b8-6219def15067] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1004.946596] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a541d296-65a4-4731-9569-74b8f931cfc2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.954272] env[62385]: DEBUG oslo_vmware.api [None req-c637a0de-3a26-483f-8836-57b0fe6d6848 tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Waiting for the task: (returnval){ [ 1004.954272] env[62385]: value = "task-1206315" [ 1004.954272] env[62385]: _type = "Task" [ 1004.954272] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.963586] env[62385]: DEBUG oslo_vmware.api [None req-c637a0de-3a26-483f-8836-57b0fe6d6848 tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Task: {'id': task-1206315, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.050924] env[62385]: DEBUG nova.compute.manager [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1005.056268] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3e96c69-cc9a-4652-b70b-61f90448ce06 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.069748] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d10603f-13f0-4fcf-8961-3d3ff49e045f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.071911] env[62385]: INFO nova.compute.manager [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Took 19.99 seconds to build instance. [ 1005.102646] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84536a45-a920-415e-bd90-718d5406f229 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.114207] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81b89e87-adc1-4078-b680-24a9cdd61a7a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.121638] env[62385]: DEBUG oslo_vmware.api [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206313, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.209945} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.123512] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1005.125179] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Deleted contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1005.125179] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1005.138198] env[62385]: DEBUG nova.compute.provider_tree [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1005.167889] env[62385]: INFO nova.scheduler.client.report [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Deleted allocations for instance 8e1d416c-2624-468c-94d7-6265e9f4178f [ 1005.181202] env[62385]: DEBUG oslo_vmware.api [None req-4becbe51-f49b-4c9e-99e8-a1178b22c7f1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206314, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.315975} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.181471] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-4becbe51-f49b-4c9e-99e8-a1178b22c7f1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1005.181664] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4becbe51-f49b-4c9e-99e8-a1178b22c7f1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Deleted contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1005.181849] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4becbe51-f49b-4c9e-99e8-a1178b22c7f1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1005.182080] env[62385]: INFO nova.compute.manager [None req-4becbe51-f49b-4c9e-99e8-a1178b22c7f1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1005.182346] env[62385]: DEBUG oslo.service.loopingcall [None req-4becbe51-f49b-4c9e-99e8-a1178b22c7f1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1005.182726] env[62385]: DEBUG nova.compute.manager [-] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1005.182726] env[62385]: DEBUG nova.network.neutron [-] [instance: df020d6b-3fab-4599-a342-47c7833b4240] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1005.221743] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9eff707c-16e0-46c5-b85d-791396fd0db5 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "5daf0a99-0c2a-40d8-afc7-1998e21e32d8" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.361s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.442406] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1005.442721] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1005.464008] env[62385]: DEBUG oslo_vmware.api [None req-c637a0de-3a26-483f-8836-57b0fe6d6848 tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Task: {'id': task-1206315, 'name': PowerOffVM_Task, 'duration_secs': 0.372636} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.465220] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c637a0de-3a26-483f-8836-57b0fe6d6848 tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] [instance: 784589bb-a668-4674-83b8-6219def15067] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1005.465409] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c637a0de-3a26-483f-8836-57b0fe6d6848 tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] [instance: 784589bb-a668-4674-83b8-6219def15067] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1005.465674] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-42d4c297-f747-4599-8904-edad151506d1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.545324] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c637a0de-3a26-483f-8836-57b0fe6d6848 tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] [instance: 784589bb-a668-4674-83b8-6219def15067] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1005.545823] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c637a0de-3a26-483f-8836-57b0fe6d6848 tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] [instance: 784589bb-a668-4674-83b8-6219def15067] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1005.545823] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c637a0de-3a26-483f-8836-57b0fe6d6848 tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Deleting the datastore file [datastore1] 784589bb-a668-4674-83b8-6219def15067 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1005.546096] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-74f29dbb-b715-493e-9e79-c2c11e804e14 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.559089] env[62385]: DEBUG oslo_vmware.api [None req-c637a0de-3a26-483f-8836-57b0fe6d6848 tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Waiting for the task: (returnval){ [ 1005.559089] env[62385]: value = "task-1206317" [ 1005.559089] env[62385]: _type = "Task" [ 1005.559089] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.573636] env[62385]: DEBUG oslo_concurrency.lockutils [None req-050b7810-165b-4433-b567-cf509b51622a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "57637d6e-39e2-405c-bbcf-64bf0ffbebd1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.498s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.573874] env[62385]: DEBUG oslo_vmware.api [None req-c637a0de-3a26-483f-8836-57b0fe6d6848 tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Task: {'id': task-1206317, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.574849] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.640421] env[62385]: INFO nova.compute.manager [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Rebuilding instance [ 1005.644733] env[62385]: DEBUG nova.scheduler.client.report [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1005.672472] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.688084] env[62385]: DEBUG nova.compute.manager [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1005.688706] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-254f8a8f-f739-41b2-93b4-99a687e1c824 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.948600] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "refresh_cache-9b9fd167-d843-4b9d-9e4f-dbde5d3d629e" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.948782] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquired lock "refresh_cache-9b9fd167-d843-4b9d-9e4f-dbde5d3d629e" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.949177] env[62385]: DEBUG nova.network.neutron [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1005.950214] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1005.950449] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Starting heal instance info cache {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1006.074186] env[62385]: DEBUG oslo_vmware.api [None req-c637a0de-3a26-483f-8836-57b0fe6d6848 tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Task: {'id': task-1206317, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.150920] env[62385]: DEBUG oslo_concurrency.lockutils [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.627s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.153803] env[62385]: DEBUG oslo_concurrency.lockutils [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 1.317s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.176866] env[62385]: INFO nova.scheduler.client.report [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Deleted allocations for instance 76672404-2bcf-438a-9e21-92ea9dc86461 [ 1006.204574] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1006.205870] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-53569ca0-cf98-4398-8532-c95cc7cfadac {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.214926] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1006.214926] env[62385]: value = "task-1206319" [ 1006.214926] env[62385]: _type = "Task" [ 1006.214926] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.230926] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206319, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.420082] env[62385]: DEBUG nova.network.neutron [-] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.572716] env[62385]: DEBUG oslo_vmware.api [None req-c637a0de-3a26-483f-8836-57b0fe6d6848 tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Task: {'id': task-1206317, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.607191} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.572995] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c637a0de-3a26-483f-8836-57b0fe6d6848 tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1006.573209] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c637a0de-3a26-483f-8836-57b0fe6d6848 tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] [instance: 784589bb-a668-4674-83b8-6219def15067] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1006.573443] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c637a0de-3a26-483f-8836-57b0fe6d6848 tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] [instance: 784589bb-a668-4674-83b8-6219def15067] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1006.573574] env[62385]: INFO nova.compute.manager [None req-c637a0de-3a26-483f-8836-57b0fe6d6848 tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] [instance: 784589bb-a668-4674-83b8-6219def15067] Took 1.64 seconds to destroy the instance on the hypervisor. [ 1006.573817] env[62385]: DEBUG oslo.service.loopingcall [None req-c637a0de-3a26-483f-8836-57b0fe6d6848 tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1006.574048] env[62385]: DEBUG nova.compute.manager [-] [instance: 784589bb-a668-4674-83b8-6219def15067] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1006.574155] env[62385]: DEBUG nova.network.neutron [-] [instance: 784589bb-a668-4674-83b8-6219def15067] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1006.683749] env[62385]: DEBUG nova.network.neutron [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Updating instance_info_cache with network_info: [{"id": "404c104c-258f-480e-bfe2-4ace4e83a43c", "address": "fa:16:3e:87:f8:2a", "network": {"id": "06017161-984c-4657-9639-4e6bdb6d4c09", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1072882119-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.208", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d45b5346eab44707b4d81eaaf12b0935", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39cd75b0-9ec7-48ed-b57f-34da0c573a60", "external-id": "nsx-vlan-transportzone-751", "segmentation_id": 751, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap404c104c-25", "ovs_interfaceid": "404c104c-258f-480e-bfe2-4ace4e83a43c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.689367] env[62385]: DEBUG oslo_concurrency.lockutils [None req-eaccdf87-7891-4bd2-b052-01a00f725c84 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "76672404-2bcf-438a-9e21-92ea9dc86461" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.144s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.728247] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206319, 'name': PowerOffVM_Task, 'duration_secs': 0.256864} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.728247] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1006.789813] env[62385]: INFO nova.compute.manager [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Detaching volume 9621f12c-980a-4762-8ea2-c4fff11dc20e [ 1006.825346] env[62385]: INFO nova.virt.block_device [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Attempting to driver detach volume 9621f12c-980a-4762-8ea2-c4fff11dc20e from mountpoint /dev/sdb [ 1006.825976] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Volume detach. Driver type: vmdk {{(pid=62385) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1006.826290] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261287', 'volume_id': '9621f12c-980a-4762-8ea2-c4fff11dc20e', 'name': 'volume-9621f12c-980a-4762-8ea2-c4fff11dc20e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5daf0a99-0c2a-40d8-afc7-1998e21e32d8', 'attached_at': '', 'detached_at': '', 'volume_id': '9621f12c-980a-4762-8ea2-c4fff11dc20e', 'serial': '9621f12c-980a-4762-8ea2-c4fff11dc20e'} {{(pid=62385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1006.827633] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b02529ed-c10a-4c10-9ab2-1584bcbc4f6a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.860470] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3fd750f-d248-4391-bd54-db22dc992531 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.863700] env[62385]: DEBUG nova.compute.manager [req-17c76b15-ca70-4f0f-a804-e7fd382f7295 req-ab1a70ca-1eb2-45ff-bb27-a62a150ae9aa service nova] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Received event network-changed-deafdeb9-f59c-48f4-873b-6e4b27d4a688 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1006.863891] env[62385]: DEBUG nova.compute.manager [req-17c76b15-ca70-4f0f-a804-e7fd382f7295 req-ab1a70ca-1eb2-45ff-bb27-a62a150ae9aa service nova] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Refreshing instance network info cache due to event network-changed-deafdeb9-f59c-48f4-873b-6e4b27d4a688. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1006.864154] env[62385]: DEBUG oslo_concurrency.lockutils [req-17c76b15-ca70-4f0f-a804-e7fd382f7295 req-ab1a70ca-1eb2-45ff-bb27-a62a150ae9aa service nova] Acquiring lock "refresh_cache-8e1d416c-2624-468c-94d7-6265e9f4178f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.864830] env[62385]: DEBUG oslo_concurrency.lockutils [req-17c76b15-ca70-4f0f-a804-e7fd382f7295 req-ab1a70ca-1eb2-45ff-bb27-a62a150ae9aa service nova] Acquired lock "refresh_cache-8e1d416c-2624-468c-94d7-6265e9f4178f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.864830] env[62385]: DEBUG nova.network.neutron [req-17c76b15-ca70-4f0f-a804-e7fd382f7295 req-ab1a70ca-1eb2-45ff-bb27-a62a150ae9aa service nova] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Refreshing network info cache for port deafdeb9-f59c-48f4-873b-6e4b27d4a688 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1006.875254] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4192094-5188-4cc8-b549-29fd5e7f79d0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.879632] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c2a0d8e-92d9-43b8-ad92-4ce01c86677d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.905137] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99eefa93-528f-4e23-9685-cc7e4bfd1279 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.909643] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f499d60-d3d3-4867-8c9e-8a6e69231bfc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.928075] env[62385]: INFO nova.compute.manager [-] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Took 1.75 seconds to deallocate network for instance. [ 1006.928460] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] The volume has not been displaced from its original location: [datastore1] volume-9621f12c-980a-4762-8ea2-c4fff11dc20e/volume-9621f12c-980a-4762-8ea2-c4fff11dc20e.vmdk. No consolidation needed. {{(pid=62385) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1006.933885] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Reconfiguring VM instance instance-0000005d to detach disk 2001 {{(pid=62385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1006.960126] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9f5c902c-f221-4c08-8dd8-74666a0aaa3e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.977931] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2441e48b-ede8-432c-8e11-5eba84650882 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.989966] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adb90db0-1407-4fa2-b982-7870e7f81851 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.994208] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1006.994208] env[62385]: value = "task-1206320" [ 1006.994208] env[62385]: _type = "Task" [ 1006.994208] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.009150] env[62385]: DEBUG nova.compute.provider_tree [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1007.014660] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206320, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.144817] env[62385]: DEBUG oslo_concurrency.lockutils [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "68748d25-017f-4b58-9984-70264abbc5ba" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.145089] env[62385]: DEBUG oslo_concurrency.lockutils [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "68748d25-017f-4b58-9984-70264abbc5ba" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.186342] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Releasing lock "refresh_cache-9b9fd167-d843-4b9d-9e4f-dbde5d3d629e" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.373734] env[62385]: DEBUG nova.network.neutron [-] [instance: 784589bb-a668-4674-83b8-6219def15067] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.483750] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4becbe51-f49b-4c9e-99e8-a1178b22c7f1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.507426] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206320, 'name': ReconfigVM_Task, 'duration_secs': 0.215636} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.507426] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Reconfigured VM instance instance-0000005d to detach disk 2001 {{(pid=62385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1007.513618] env[62385]: DEBUG nova.scheduler.client.report [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1007.516816] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fe7d01f3-6622-47dd-bb42-6f7f6efddf81 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.538599] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1007.538599] env[62385]: value = "task-1206321" [ 1007.538599] env[62385]: _type = "Task" [ 1007.538599] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.550613] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206321, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.648163] env[62385]: DEBUG nova.compute.manager [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1007.709608] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb55306d-ea42-4cc7-8e34-563639b45271 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.713481] env[62385]: DEBUG nova.network.neutron [req-17c76b15-ca70-4f0f-a804-e7fd382f7295 req-ab1a70ca-1eb2-45ff-bb27-a62a150ae9aa service nova] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Updated VIF entry in instance network info cache for port deafdeb9-f59c-48f4-873b-6e4b27d4a688. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1007.713825] env[62385]: DEBUG nova.network.neutron [req-17c76b15-ca70-4f0f-a804-e7fd382f7295 req-ab1a70ca-1eb2-45ff-bb27-a62a150ae9aa service nova] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Updating instance_info_cache with network_info: [{"id": "deafdeb9-f59c-48f4-873b-6e4b27d4a688", "address": "fa:16:3e:81:4d:75", "network": {"id": "9c4d6464-cacc-44f8-b7e5-0699cd312c53", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1889153615-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.186", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09c2c5e938bf46fd9353e47aecc72478", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapdeafdeb9-f5", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.731655] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a04c0d19-d087-4256-8206-6f67348c61d6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.740675] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Updating instance '9b9fd167-d843-4b9d-9e4f-dbde5d3d629e' progress to 83 {{(pid=62385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1007.876723] env[62385]: INFO nova.compute.manager [-] [instance: 784589bb-a668-4674-83b8-6219def15067] Took 1.30 seconds to deallocate network for instance. [ 1008.034221] env[62385]: DEBUG oslo_concurrency.lockutils [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 1.880s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.036914] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.462s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.038426] env[62385]: INFO nova.compute.claims [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1008.051890] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206321, 'name': ReconfigVM_Task, 'duration_secs': 0.365265} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.052357] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261287', 'volume_id': '9621f12c-980a-4762-8ea2-c4fff11dc20e', 'name': 'volume-9621f12c-980a-4762-8ea2-c4fff11dc20e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5daf0a99-0c2a-40d8-afc7-1998e21e32d8', 'attached_at': '', 'detached_at': '', 'volume_id': '9621f12c-980a-4762-8ea2-c4fff11dc20e', 'serial': '9621f12c-980a-4762-8ea2-c4fff11dc20e'} {{(pid=62385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1008.170300] env[62385]: DEBUG oslo_concurrency.lockutils [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.231065] env[62385]: DEBUG oslo_concurrency.lockutils [req-17c76b15-ca70-4f0f-a804-e7fd382f7295 req-ab1a70ca-1eb2-45ff-bb27-a62a150ae9aa service nova] Releasing lock "refresh_cache-8e1d416c-2624-468c-94d7-6265e9f4178f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1008.231374] env[62385]: DEBUG nova.compute.manager [req-17c76b15-ca70-4f0f-a804-e7fd382f7295 req-ab1a70ca-1eb2-45ff-bb27-a62a150ae9aa service nova] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Received event network-vif-deleted-a9513285-4023-4f3e-b90d-3eb507955f13 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1008.231570] env[62385]: DEBUG nova.compute.manager [req-17c76b15-ca70-4f0f-a804-e7fd382f7295 req-ab1a70ca-1eb2-45ff-bb27-a62a150ae9aa service nova] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Received event network-changed-a7153a2d-db69-4d14-8ddc-b634ddabbc18 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1008.231737] env[62385]: DEBUG nova.compute.manager [req-17c76b15-ca70-4f0f-a804-e7fd382f7295 req-ab1a70ca-1eb2-45ff-bb27-a62a150ae9aa service nova] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Refreshing instance network info cache due to event network-changed-a7153a2d-db69-4d14-8ddc-b634ddabbc18. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1008.231965] env[62385]: DEBUG oslo_concurrency.lockutils [req-17c76b15-ca70-4f0f-a804-e7fd382f7295 req-ab1a70ca-1eb2-45ff-bb27-a62a150ae9aa service nova] Acquiring lock "refresh_cache-57637d6e-39e2-405c-bbcf-64bf0ffbebd1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.232131] env[62385]: DEBUG oslo_concurrency.lockutils [req-17c76b15-ca70-4f0f-a804-e7fd382f7295 req-ab1a70ca-1eb2-45ff-bb27-a62a150ae9aa service nova] Acquired lock "refresh_cache-57637d6e-39e2-405c-bbcf-64bf0ffbebd1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.232299] env[62385]: DEBUG nova.network.neutron [req-17c76b15-ca70-4f0f-a804-e7fd382f7295 req-ab1a70ca-1eb2-45ff-bb27-a62a150ae9aa service nova] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Refreshing network info cache for port a7153a2d-db69-4d14-8ddc-b634ddabbc18 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1008.246912] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1008.247219] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-da84492d-4d29-4a2d-8e06-af201313a098 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.254650] env[62385]: DEBUG oslo_vmware.api [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1008.254650] env[62385]: value = "task-1206323" [ 1008.254650] env[62385]: _type = "Task" [ 1008.254650] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.263164] env[62385]: DEBUG oslo_vmware.api [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206323, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.264651] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquiring lock "8e1d416c-2624-468c-94d7-6265e9f4178f" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.385349] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c637a0de-3a26-483f-8836-57b0fe6d6848 tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.424798] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "47311a9c-74d2-400d-a22d-9ef55b14edb5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.424798] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "47311a9c-74d2-400d-a22d-9ef55b14edb5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.599932] env[62385]: INFO nova.scheduler.client.report [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Deleted allocation for migration 103f026b-ef52-4238-8a48-c1a1b0aed1b6 [ 1008.764961] env[62385]: DEBUG oslo_vmware.api [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206323, 'name': PowerOnVM_Task, 'duration_secs': 0.414797} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.765312] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1008.765458] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4c04022e-4553-4c9c-ad70-d08b4757d8bc tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Updating instance '9b9fd167-d843-4b9d-9e4f-dbde5d3d629e' progress to 100 {{(pid=62385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1008.928200] env[62385]: DEBUG nova.compute.manager [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1008.933309] env[62385]: DEBUG nova.compute.manager [req-8ed8d9be-df50-480d-8b22-b5cff960eed2 req-9c9d4c01-0ab8-4a7d-b19e-38726f0d1906 service nova] [instance: 784589bb-a668-4674-83b8-6219def15067] Received event network-vif-deleted-682074de-7041-442e-bc00-fcefe591c878 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1008.965787] env[62385]: DEBUG nova.network.neutron [req-17c76b15-ca70-4f0f-a804-e7fd382f7295 req-ab1a70ca-1eb2-45ff-bb27-a62a150ae9aa service nova] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Updated VIF entry in instance network info cache for port a7153a2d-db69-4d14-8ddc-b634ddabbc18. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1008.966137] env[62385]: DEBUG nova.network.neutron [req-17c76b15-ca70-4f0f-a804-e7fd382f7295 req-ab1a70ca-1eb2-45ff-bb27-a62a150ae9aa service nova] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Updating instance_info_cache with network_info: [{"id": "a7153a2d-db69-4d14-8ddc-b634ddabbc18", "address": "fa:16:3e:a3:0d:ab", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.210", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7153a2d-db", "ovs_interfaceid": "a7153a2d-db69-4d14-8ddc-b634ddabbc18", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.106610] env[62385]: DEBUG oslo_concurrency.lockutils [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "b1e0baf3-643c-49c2-8a80-8de07d8527e7" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.662s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.112352] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1009.112806] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-abff7fd2-d5e5-4621-9a35-44dc09be1a10 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.121087] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1009.121087] env[62385]: value = "task-1206324" [ 1009.121087] env[62385]: _type = "Task" [ 1009.121087] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.133134] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] VM already powered off {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1009.133983] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Volume detach. Driver type: vmdk {{(pid=62385) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1009.133983] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261287', 'volume_id': '9621f12c-980a-4762-8ea2-c4fff11dc20e', 'name': 'volume-9621f12c-980a-4762-8ea2-c4fff11dc20e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5daf0a99-0c2a-40d8-afc7-1998e21e32d8', 'attached_at': '', 'detached_at': '', 'volume_id': '9621f12c-980a-4762-8ea2-c4fff11dc20e', 'serial': '9621f12c-980a-4762-8ea2-c4fff11dc20e'} {{(pid=62385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1009.134394] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b79be579-309f-41d9-a2e9-1557647b0d99 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.158199] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a3692a5-70ab-4406-aea5-5f16d2ef7835 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.165978] env[62385]: WARNING nova.virt.vmwareapi.driver [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] The volume None does not exist!: nova.exception.DiskNotFound: Unable to find volume [ 1009.166324] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1009.167096] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d8870e0-7a47-4e0e-818b-1ba829678adb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.177516] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1009.177634] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c29947a1-f846-4e4d-95f9-b71deea996bb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.237571] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-782d9212-e768-4cfa-b627-2abd7629f6ca {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.245613] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9386cd1-b0ea-428c-a851-7b19cf12bcc7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.282232] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e3e56ab-cf3b-4f79-bc0a-194596a0a18a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.284784] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1009.285050] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1009.285265] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Deleting the datastore file [datastore1] 5daf0a99-0c2a-40d8-afc7-1998e21e32d8 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1009.285741] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-911845a5-276a-4687-8539-16a2e5a888f5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.297051] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be68e8ed-d9e7-4198-8281-8fff1e221fd3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.301406] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1009.301406] env[62385]: value = "task-1206326" [ 1009.301406] env[62385]: _type = "Task" [ 1009.301406] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.313638] env[62385]: DEBUG nova.compute.provider_tree [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1009.320568] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206326, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.450201] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.468646] env[62385]: DEBUG oslo_concurrency.lockutils [req-17c76b15-ca70-4f0f-a804-e7fd382f7295 req-ab1a70ca-1eb2-45ff-bb27-a62a150ae9aa service nova] Releasing lock "refresh_cache-57637d6e-39e2-405c-bbcf-64bf0ffbebd1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.815807] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206326, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.211542} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.816170] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1009.816398] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1009.816686] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1009.821462] env[62385]: DEBUG nova.scheduler.client.report [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1009.996339] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Didn't find any instances for network info cache update. {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1009.996910] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1009.997211] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1009.997465] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1009.997672] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1009.997866] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1009.998063] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1009.998234] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62385) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1009.998345] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager.update_available_resource {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1010.327592] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.291s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.328116] env[62385]: DEBUG nova.compute.manager [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1010.332021] env[62385]: INFO nova.virt.block_device [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Booting with volume 9621f12c-980a-4762-8ea2-c4fff11dc20e at /dev/sdb [ 1010.335595] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.661s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.335595] env[62385]: DEBUG nova.objects.instance [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lazy-loading 'resources' on Instance uuid 8e1d416c-2624-468c-94d7-6265e9f4178f {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1010.456296] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-52fc3863-25bd-49db-82d4-2c6515b74aff {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.463828] env[62385]: DEBUG oslo_concurrency.lockutils [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "b1e0baf3-643c-49c2-8a80-8de07d8527e7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.464092] env[62385]: DEBUG oslo_concurrency.lockutils [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "b1e0baf3-643c-49c2-8a80-8de07d8527e7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.464309] env[62385]: DEBUG oslo_concurrency.lockutils [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "b1e0baf3-643c-49c2-8a80-8de07d8527e7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.464496] env[62385]: DEBUG oslo_concurrency.lockutils [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "b1e0baf3-643c-49c2-8a80-8de07d8527e7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.464672] env[62385]: DEBUG oslo_concurrency.lockutils [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "b1e0baf3-643c-49c2-8a80-8de07d8527e7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.469054] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8348aaa-dadb-482c-9c27-6ca62b3b1e31 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.481436] env[62385]: INFO nova.compute.manager [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Terminating instance [ 1010.483643] env[62385]: DEBUG nova.compute.manager [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1010.483832] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1010.484581] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d610605-3c83-4976-b681-3b96cb522498 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.491265] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1010.491497] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f2c7110f-c86e-4ad6-b5ea-d19a75acb454 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.501593] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.501829] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-863b477a-d4bf-444b-9615-62868966c40a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.503600] env[62385]: DEBUG oslo_vmware.api [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 1010.503600] env[62385]: value = "task-1206328" [ 1010.503600] env[62385]: _type = "Task" [ 1010.503600] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.510744] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0194b5c0-81b9-4b7f-8c88-a67920b2ef4e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.524027] env[62385]: DEBUG oslo_vmware.api [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206328, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.541940] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-012b0675-b1a2-46e0-b094-a94e18689849 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.549090] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e43b9c0e-ea10-43bc-9734-dfdef1804857 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.563342] env[62385]: DEBUG nova.virt.block_device [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Updating existing volume attachment record: 4cfd7da2-3a7e-4eae-a550-c47d2cb64a37 {{(pid=62385) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1010.747131] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7525a56b-c44c-4f42-aefb-21ae42b9f33a tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "9b9fd167-d843-4b9d-9e4f-dbde5d3d629e" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.747424] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7525a56b-c44c-4f42-aefb-21ae42b9f33a tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "9b9fd167-d843-4b9d-9e4f-dbde5d3d629e" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.747616] env[62385]: DEBUG nova.compute.manager [None req-7525a56b-c44c-4f42-aefb-21ae42b9f33a tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Going to confirm migration 4 {{(pid=62385) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1010.835617] env[62385]: DEBUG nova.compute.utils [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1010.836478] env[62385]: DEBUG nova.compute.manager [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1010.836653] env[62385]: DEBUG nova.network.neutron [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1010.839332] env[62385]: DEBUG nova.objects.instance [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lazy-loading 'numa_topology' on Instance uuid 8e1d416c-2624-468c-94d7-6265e9f4178f {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1010.879588] env[62385]: DEBUG nova.policy [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4ac552dc01fc4de2b97e36b41dae725a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '105150c170d74e268e82ab894e9bf0d5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 1011.014434] env[62385]: DEBUG oslo_vmware.api [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206328, 'name': PowerOffVM_Task, 'duration_secs': 0.190404} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.014711] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1011.014885] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1011.015165] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-15e1ad0f-7d06-4dd0-837e-592feb4804c1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.086751] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1011.087014] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Deleting contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1011.087250] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Deleting the datastore file [datastore2] b1e0baf3-643c-49c2-8a80-8de07d8527e7 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1011.087540] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-56597330-579a-4781-a7d4-89e3a214f2fb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.094438] env[62385]: DEBUG oslo_vmware.api [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for the task: (returnval){ [ 1011.094438] env[62385]: value = "task-1206330" [ 1011.094438] env[62385]: _type = "Task" [ 1011.094438] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.102664] env[62385]: DEBUG oslo_vmware.api [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206330, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.143288] env[62385]: DEBUG nova.network.neutron [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Successfully created port: c8bba88c-566a-4a08-9526-4365303ef423 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1011.328668] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7525a56b-c44c-4f42-aefb-21ae42b9f33a tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "refresh_cache-9b9fd167-d843-4b9d-9e4f-dbde5d3d629e" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.328668] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7525a56b-c44c-4f42-aefb-21ae42b9f33a tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquired lock "refresh_cache-9b9fd167-d843-4b9d-9e4f-dbde5d3d629e" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.328668] env[62385]: DEBUG nova.network.neutron [None req-7525a56b-c44c-4f42-aefb-21ae42b9f33a tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1011.328668] env[62385]: DEBUG nova.objects.instance [None req-7525a56b-c44c-4f42-aefb-21ae42b9f33a tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lazy-loading 'info_cache' on Instance uuid 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1011.342321] env[62385]: DEBUG nova.compute.manager [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1011.346611] env[62385]: DEBUG nova.objects.base [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Object Instance<8e1d416c-2624-468c-94d7-6265e9f4178f> lazy-loaded attributes: resources,numa_topology {{(pid=62385) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1011.529811] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d95daffb-af7c-4351-b1c4-d502f7b4800a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.539239] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8c789cb-5df2-43df-b397-771f65f83f86 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.570181] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b8f6f9b-fb44-4b91-82ee-6943d93a438e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.578118] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e24950cd-c167-4683-a65d-d04851faef42 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.592355] env[62385]: DEBUG nova.compute.provider_tree [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1011.603457] env[62385]: DEBUG oslo_vmware.api [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Task: {'id': task-1206330, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.173332} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.603704] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1011.603891] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Deleted contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1011.604096] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1011.604274] env[62385]: INFO nova.compute.manager [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1011.604518] env[62385]: DEBUG oslo.service.loopingcall [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1011.604714] env[62385]: DEBUG nova.compute.manager [-] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1011.604807] env[62385]: DEBUG nova.network.neutron [-] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1011.905297] env[62385]: DEBUG nova.compute.manager [req-598df509-ac99-409b-b08c-51850ac5494c req-64a5f33d-6e31-40b8-bb47-0c9d9eb98a5e service nova] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Received event network-vif-deleted-eef524f4-fc00-47f1-a485-dcf568c6d0e4 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1011.906179] env[62385]: INFO nova.compute.manager [req-598df509-ac99-409b-b08c-51850ac5494c req-64a5f33d-6e31-40b8-bb47-0c9d9eb98a5e service nova] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Neutron deleted interface eef524f4-fc00-47f1-a485-dcf568c6d0e4; detaching it from the instance and deleting it from the info cache [ 1011.906179] env[62385]: DEBUG nova.network.neutron [req-598df509-ac99-409b-b08c-51850ac5494c req-64a5f33d-6e31-40b8-bb47-0c9d9eb98a5e service nova] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.095401] env[62385]: DEBUG nova.scheduler.client.report [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1012.358639] env[62385]: DEBUG nova.compute.manager [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1012.372130] env[62385]: DEBUG nova.network.neutron [-] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.383411] env[62385]: DEBUG nova.virt.hardware [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1012.383660] env[62385]: DEBUG nova.virt.hardware [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1012.383822] env[62385]: DEBUG nova.virt.hardware [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1012.384014] env[62385]: DEBUG nova.virt.hardware [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1012.384177] env[62385]: DEBUG nova.virt.hardware [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1012.384329] env[62385]: DEBUG nova.virt.hardware [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1012.384541] env[62385]: DEBUG nova.virt.hardware [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1012.384702] env[62385]: DEBUG nova.virt.hardware [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1012.384871] env[62385]: DEBUG nova.virt.hardware [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1012.385046] env[62385]: DEBUG nova.virt.hardware [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1012.385228] env[62385]: DEBUG nova.virt.hardware [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1012.386348] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad3f2239-70b3-4660-ae4e-b2029c8fc2ba {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.395366] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d335efdf-d512-4f97-a9fc-a7af15eecc12 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.409601] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b97ba39c-695b-47fb-8368-9ab1eb955274 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.418052] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40d6c58e-dbfd-495b-889f-61e6d0ff5f34 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.446396] env[62385]: DEBUG nova.compute.manager [req-598df509-ac99-409b-b08c-51850ac5494c req-64a5f33d-6e31-40b8-bb47-0c9d9eb98a5e service nova] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Detach interface failed, port_id=eef524f4-fc00-47f1-a485-dcf568c6d0e4, reason: Instance b1e0baf3-643c-49c2-8a80-8de07d8527e7 could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1012.581813] env[62385]: DEBUG nova.network.neutron [None req-7525a56b-c44c-4f42-aefb-21ae42b9f33a tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Updating instance_info_cache with network_info: [{"id": "404c104c-258f-480e-bfe2-4ace4e83a43c", "address": "fa:16:3e:87:f8:2a", "network": {"id": "06017161-984c-4657-9639-4e6bdb6d4c09", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1072882119-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.208", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d45b5346eab44707b4d81eaaf12b0935", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39cd75b0-9ec7-48ed-b57f-34da0c573a60", "external-id": "nsx-vlan-transportzone-751", "segmentation_id": 751, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap404c104c-25", "ovs_interfaceid": "404c104c-258f-480e-bfe2-4ace4e83a43c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.600955] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.267s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.603295] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4becbe51-f49b-4c9e-99e8-a1178b22c7f1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.120s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.603489] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4becbe51-f49b-4c9e-99e8-a1178b22c7f1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.605597] env[62385]: DEBUG oslo_concurrency.lockutils [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.435s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.607385] env[62385]: INFO nova.compute.claims [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1012.632306] env[62385]: INFO nova.scheduler.client.report [None req-4becbe51-f49b-4c9e-99e8-a1178b22c7f1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Deleted allocations for instance df020d6b-3fab-4599-a342-47c7833b4240 [ 1012.670930] env[62385]: DEBUG nova.virt.hardware [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1012.671162] env[62385]: DEBUG nova.virt.hardware [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1012.671328] env[62385]: DEBUG nova.virt.hardware [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1012.671566] env[62385]: DEBUG nova.virt.hardware [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1012.671655] env[62385]: DEBUG nova.virt.hardware [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1012.671803] env[62385]: DEBUG nova.virt.hardware [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1012.672021] env[62385]: DEBUG nova.virt.hardware [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1012.672203] env[62385]: DEBUG nova.virt.hardware [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1012.672463] env[62385]: DEBUG nova.virt.hardware [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1012.672534] env[62385]: DEBUG nova.virt.hardware [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1012.672716] env[62385]: DEBUG nova.virt.hardware [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1012.673808] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86f9e1a1-a26d-4e59-bb41-b23b4d0fb5cb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.682117] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-301cac81-deac-4ff5-8e3d-2015ca521dd3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.697811] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:85:9d:e7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4adc8ed0-d11a-4510-9be0-b27c0da3a903', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1b04305d-1213-4f55-9b69-6f15de71f576', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1012.705267] env[62385]: DEBUG oslo.service.loopingcall [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1012.705515] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1012.705958] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-eff5b833-18eb-421c-bf0a-27473c9a9a04 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.727007] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1012.727007] env[62385]: value = "task-1206331" [ 1012.727007] env[62385]: _type = "Task" [ 1012.727007] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.738593] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206331, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.876580] env[62385]: INFO nova.compute.manager [-] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Took 1.27 seconds to deallocate network for instance. [ 1013.006776] env[62385]: DEBUG nova.compute.manager [req-1ceb49e5-147c-4f55-a8bf-81f0d4a3111a req-77a411e5-c32c-44aa-bde3-f41cf57696bb service nova] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Received event network-vif-plugged-c8bba88c-566a-4a08-9526-4365303ef423 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1013.007093] env[62385]: DEBUG oslo_concurrency.lockutils [req-1ceb49e5-147c-4f55-a8bf-81f0d4a3111a req-77a411e5-c32c-44aa-bde3-f41cf57696bb service nova] Acquiring lock "b0a6a3b9-05e4-4190-9cc6-4237dee07f91-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.007232] env[62385]: DEBUG oslo_concurrency.lockutils [req-1ceb49e5-147c-4f55-a8bf-81f0d4a3111a req-77a411e5-c32c-44aa-bde3-f41cf57696bb service nova] Lock "b0a6a3b9-05e4-4190-9cc6-4237dee07f91-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.007408] env[62385]: DEBUG oslo_concurrency.lockutils [req-1ceb49e5-147c-4f55-a8bf-81f0d4a3111a req-77a411e5-c32c-44aa-bde3-f41cf57696bb service nova] Lock "b0a6a3b9-05e4-4190-9cc6-4237dee07f91-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.007688] env[62385]: DEBUG nova.compute.manager [req-1ceb49e5-147c-4f55-a8bf-81f0d4a3111a req-77a411e5-c32c-44aa-bde3-f41cf57696bb service nova] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] No waiting events found dispatching network-vif-plugged-c8bba88c-566a-4a08-9526-4365303ef423 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1013.007824] env[62385]: WARNING nova.compute.manager [req-1ceb49e5-147c-4f55-a8bf-81f0d4a3111a req-77a411e5-c32c-44aa-bde3-f41cf57696bb service nova] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Received unexpected event network-vif-plugged-c8bba88c-566a-4a08-9526-4365303ef423 for instance with vm_state building and task_state spawning. [ 1013.031851] env[62385]: DEBUG nova.network.neutron [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Successfully updated port: c8bba88c-566a-4a08-9526-4365303ef423 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1013.085505] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7525a56b-c44c-4f42-aefb-21ae42b9f33a tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Releasing lock "refresh_cache-9b9fd167-d843-4b9d-9e4f-dbde5d3d629e" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.085779] env[62385]: DEBUG nova.objects.instance [None req-7525a56b-c44c-4f42-aefb-21ae42b9f33a tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lazy-loading 'migration_context' on Instance uuid 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1013.117182] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3ce49b57-206e-45d9-8db4-b5ed524447f3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lock "8e1d416c-2624-468c-94d7-6265e9f4178f" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 27.815s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.117182] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lock "8e1d416c-2624-468c-94d7-6265e9f4178f" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 4.852s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.117182] env[62385]: INFO nova.compute.manager [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Unshelving [ 1013.140181] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4becbe51-f49b-4c9e-99e8-a1178b22c7f1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "df020d6b-3fab-4599-a342-47c7833b4240" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.102s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.238267] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206331, 'name': CreateVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.384352] env[62385]: DEBUG oslo_concurrency.lockutils [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.535092] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "refresh_cache-b0a6a3b9-05e4-4190-9cc6-4237dee07f91" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.535092] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquired lock "refresh_cache-b0a6a3b9-05e4-4190-9cc6-4237dee07f91" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.535092] env[62385]: DEBUG nova.network.neutron [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1013.591699] env[62385]: DEBUG nova.objects.base [None req-7525a56b-c44c-4f42-aefb-21ae42b9f33a tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Object Instance<9b9fd167-d843-4b9d-9e4f-dbde5d3d629e> lazy-loaded attributes: info_cache,migration_context {{(pid=62385) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1013.592729] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-833060fe-c9f7-4398-b750-0cc2dd67c7f0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.612342] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1dc2609-31af-4686-a01f-21bebdfafa0d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.620624] env[62385]: DEBUG oslo_vmware.api [None req-7525a56b-c44c-4f42-aefb-21ae42b9f33a tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1013.620624] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]527688d3-d94c-5a21-fa2d-69f376ef88c6" [ 1013.620624] env[62385]: _type = "Task" [ 1013.620624] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.631358] env[62385]: DEBUG oslo_vmware.api [None req-7525a56b-c44c-4f42-aefb-21ae42b9f33a tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]527688d3-d94c-5a21-fa2d-69f376ef88c6, 'name': SearchDatastore_Task, 'duration_secs': 0.009357} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.631524] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7525a56b-c44c-4f42-aefb-21ae42b9f33a tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.657305] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Acquiring lock "5a5bcb7c-86cd-4950-a3dd-15515ae0aca6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.657610] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Lock "5a5bcb7c-86cd-4950-a3dd-15515ae0aca6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.738941] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206331, 'name': CreateVM_Task, 'duration_secs': 0.612062} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.741193] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1013.742043] env[62385]: DEBUG oslo_concurrency.lockutils [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.742267] env[62385]: DEBUG oslo_concurrency.lockutils [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.742596] env[62385]: DEBUG oslo_concurrency.lockutils [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1013.742861] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-566952cf-08dc-4c20-8226-eb9d5230ad53 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.747614] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1013.747614] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5238fdf2-5bd4-eb94-13ce-c7d2cc5e98d7" [ 1013.747614] env[62385]: _type = "Task" [ 1013.747614] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.758845] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5238fdf2-5bd4-eb94-13ce-c7d2cc5e98d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.788066] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd1547dd-7ee2-45e6-a89f-7bc7cd85b46c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.798405] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-499ee660-72d5-4f9d-98db-a41b816146a5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.847760] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4180e786-656d-428a-a852-e6b5a6dcfc92 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.858238] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-526c6143-f181-45bf-be22-0eb508f711de {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.879139] env[62385]: DEBUG nova.compute.provider_tree [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1014.084822] env[62385]: DEBUG nova.network.neutron [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1014.133073] env[62385]: DEBUG nova.compute.utils [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1014.165584] env[62385]: DEBUG nova.compute.manager [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1014.257994] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5238fdf2-5bd4-eb94-13ce-c7d2cc5e98d7, 'name': SearchDatastore_Task, 'duration_secs': 0.011587} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.258725] env[62385]: DEBUG oslo_concurrency.lockutils [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.259056] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1014.259531] env[62385]: DEBUG oslo_concurrency.lockutils [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.259718] env[62385]: DEBUG oslo_concurrency.lockutils [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.260019] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1014.260351] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8b58e686-e782-4b43-9de3-730d2191bf74 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.270265] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1014.270494] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1014.271247] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e2ce973-454b-4d8a-83d7-31b6e063de47 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.277254] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1014.277254] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]527c5a3e-c9d8-c1c8-fa85-bccba6cf9117" [ 1014.277254] env[62385]: _type = "Task" [ 1014.277254] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.285838] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]527c5a3e-c9d8-c1c8-fa85-bccba6cf9117, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.325085] env[62385]: DEBUG nova.network.neutron [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Updating instance_info_cache with network_info: [{"id": "c8bba88c-566a-4a08-9526-4365303ef423", "address": "fa:16:3e:50:87:ae", "network": {"id": "30c4e78b-6a64-47dc-aaaa-a210160c9f23", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1809587656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "105150c170d74e268e82ab894e9bf0d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10ff2092-e8eb-4768-ad4a-65a80560b447", "external-id": "nsx-vlan-transportzone-845", "segmentation_id": 845, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8bba88c-56", "ovs_interfaceid": "c8bba88c-566a-4a08-9526-4365303ef423", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.383225] env[62385]: DEBUG nova.scheduler.client.report [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1014.638285] env[62385]: INFO nova.virt.block_device [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Booting with volume 925fe844-0d8f-4668-ac82-328339e88cac at /dev/sdb [ 1014.678046] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c71055ea-bdee-4be8-b75a-29cc33072d64 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.687520] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebd155ef-dd21-49de-bbca-20b36a1d9832 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.699209] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.717921] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e702592c-03a7-4367-ab1e-6ab97723421a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.727156] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5a803aa-4d66-4457-9d55-4bf0778f45e2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.757827] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a2a6a82-60a8-4a87-bc30-dac3244242b6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.765173] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cb7a986-6c40-46d5-ae29-0059c83a64b5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.778989] env[62385]: DEBUG nova.virt.block_device [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Updating existing volume attachment record: d3b7c0c2-0351-4826-ad5c-f584c6d8097c {{(pid=62385) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1014.790209] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]527c5a3e-c9d8-c1c8-fa85-bccba6cf9117, 'name': SearchDatastore_Task, 'duration_secs': 0.00905} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.791068] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c03bda2b-eebb-4b78-9e34-d28007edfd7c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.796758] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1014.796758] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5272112d-c653-fe9f-d21e-b21f4fa47172" [ 1014.796758] env[62385]: _type = "Task" [ 1014.796758] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.806854] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5272112d-c653-fe9f-d21e-b21f4fa47172, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.828550] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Releasing lock "refresh_cache-b0a6a3b9-05e4-4190-9cc6-4237dee07f91" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.828907] env[62385]: DEBUG nova.compute.manager [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Instance network_info: |[{"id": "c8bba88c-566a-4a08-9526-4365303ef423", "address": "fa:16:3e:50:87:ae", "network": {"id": "30c4e78b-6a64-47dc-aaaa-a210160c9f23", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1809587656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "105150c170d74e268e82ab894e9bf0d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10ff2092-e8eb-4768-ad4a-65a80560b447", "external-id": "nsx-vlan-transportzone-845", "segmentation_id": 845, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8bba88c-56", "ovs_interfaceid": "c8bba88c-566a-4a08-9526-4365303ef423", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1014.829327] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:50:87:ae', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '10ff2092-e8eb-4768-ad4a-65a80560b447', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c8bba88c-566a-4a08-9526-4365303ef423', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1014.836814] env[62385]: DEBUG oslo.service.loopingcall [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1014.837050] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1014.837284] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9522ae49-2e66-4a9d-a12e-d668e7ee48bf {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.856603] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1014.856603] env[62385]: value = "task-1206332" [ 1014.856603] env[62385]: _type = "Task" [ 1014.856603] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.864374] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206332, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.888557] env[62385]: DEBUG oslo_concurrency.lockutils [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.283s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.889104] env[62385]: DEBUG nova.compute.manager [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1014.892208] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c637a0de-3a26-483f-8836-57b0fe6d6848 tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.507s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.892348] env[62385]: DEBUG nova.objects.instance [None req-c637a0de-3a26-483f-8836-57b0fe6d6848 tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Lazy-loading 'resources' on Instance uuid 784589bb-a668-4674-83b8-6219def15067 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1015.032599] env[62385]: DEBUG nova.compute.manager [req-0ca9725b-3fbd-48ae-b282-b6f57fb258a5 req-7596b8b9-e5d3-41bc-bffd-230c62d0c891 service nova] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Received event network-changed-c8bba88c-566a-4a08-9526-4365303ef423 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1015.032970] env[62385]: DEBUG nova.compute.manager [req-0ca9725b-3fbd-48ae-b282-b6f57fb258a5 req-7596b8b9-e5d3-41bc-bffd-230c62d0c891 service nova] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Refreshing instance network info cache due to event network-changed-c8bba88c-566a-4a08-9526-4365303ef423. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1015.033235] env[62385]: DEBUG oslo_concurrency.lockutils [req-0ca9725b-3fbd-48ae-b282-b6f57fb258a5 req-7596b8b9-e5d3-41bc-bffd-230c62d0c891 service nova] Acquiring lock "refresh_cache-b0a6a3b9-05e4-4190-9cc6-4237dee07f91" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.033387] env[62385]: DEBUG oslo_concurrency.lockutils [req-0ca9725b-3fbd-48ae-b282-b6f57fb258a5 req-7596b8b9-e5d3-41bc-bffd-230c62d0c891 service nova] Acquired lock "refresh_cache-b0a6a3b9-05e4-4190-9cc6-4237dee07f91" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.033556] env[62385]: DEBUG nova.network.neutron [req-0ca9725b-3fbd-48ae-b282-b6f57fb258a5 req-7596b8b9-e5d3-41bc-bffd-230c62d0c891 service nova] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Refreshing network info cache for port c8bba88c-566a-4a08-9526-4365303ef423 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1015.307677] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5272112d-c653-fe9f-d21e-b21f4fa47172, 'name': SearchDatastore_Task, 'duration_secs': 0.010562} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.308139] env[62385]: DEBUG oslo_concurrency.lockutils [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.308230] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 5daf0a99-0c2a-40d8-afc7-1998e21e32d8/5daf0a99-0c2a-40d8-afc7-1998e21e32d8.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1015.308488] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e082bda8-f9f6-4b56-addb-31b025c7e005 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.315658] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1015.315658] env[62385]: value = "task-1206336" [ 1015.315658] env[62385]: _type = "Task" [ 1015.315658] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.325243] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206336, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.367707] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206332, 'name': CreateVM_Task, 'duration_secs': 0.353131} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.367950] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1015.368821] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.369089] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.369514] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1015.369832] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4893a85b-273f-4d5b-ad18-d2572b0926d3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.375454] env[62385]: DEBUG oslo_vmware.api [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 1015.375454] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ba4150-a026-f20f-1f4f-71641c680b47" [ 1015.375454] env[62385]: _type = "Task" [ 1015.375454] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.384447] env[62385]: DEBUG oslo_vmware.api [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ba4150-a026-f20f-1f4f-71641c680b47, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.396139] env[62385]: DEBUG nova.compute.utils [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1015.401019] env[62385]: DEBUG nova.compute.manager [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1015.401019] env[62385]: DEBUG nova.network.neutron [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1015.477466] env[62385]: DEBUG nova.policy [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5040b20c848c4d28a198f1773a3ff21e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '012426f098ce40c3aaa00f628fe9cebb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 1015.590253] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2054029-21a5-4240-8168-e6a858319412 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.601744] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-728a1b38-d403-4a48-a4f4-19f3f4bbfc3e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.636103] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73e0ba0e-3148-4aad-9ae2-2b25212c44dc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.645565] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-184ef92e-02a7-41c3-9fbe-902cfa0c1767 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.660633] env[62385]: DEBUG nova.compute.provider_tree [None req-c637a0de-3a26-483f-8836-57b0fe6d6848 tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1015.822072] env[62385]: DEBUG nova.network.neutron [req-0ca9725b-3fbd-48ae-b282-b6f57fb258a5 req-7596b8b9-e5d3-41bc-bffd-230c62d0c891 service nova] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Updated VIF entry in instance network info cache for port c8bba88c-566a-4a08-9526-4365303ef423. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1015.822723] env[62385]: DEBUG nova.network.neutron [req-0ca9725b-3fbd-48ae-b282-b6f57fb258a5 req-7596b8b9-e5d3-41bc-bffd-230c62d0c891 service nova] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Updating instance_info_cache with network_info: [{"id": "c8bba88c-566a-4a08-9526-4365303ef423", "address": "fa:16:3e:50:87:ae", "network": {"id": "30c4e78b-6a64-47dc-aaaa-a210160c9f23", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1809587656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "105150c170d74e268e82ab894e9bf0d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10ff2092-e8eb-4768-ad4a-65a80560b447", "external-id": "nsx-vlan-transportzone-845", "segmentation_id": 845, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8bba88c-56", "ovs_interfaceid": "c8bba88c-566a-4a08-9526-4365303ef423", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.831395] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206336, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.882910] env[62385]: DEBUG nova.network.neutron [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Successfully created port: c308701b-9b9e-4ff3-9a01-c27aabf54f0d {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1015.891149] env[62385]: DEBUG oslo_vmware.api [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ba4150-a026-f20f-1f4f-71641c680b47, 'name': SearchDatastore_Task, 'duration_secs': 0.013281} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.891480] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.891816] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1015.891950] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.892116] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.892327] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1015.892603] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9f31194a-082b-4a43-acca-9277fd914f84 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.902204] env[62385]: DEBUG nova.compute.manager [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1015.904963] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1015.905075] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1015.906269] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9b57d15-ce24-4750-b198-e0381aac7588 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.912986] env[62385]: DEBUG oslo_vmware.api [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 1015.912986] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c19ca8-f566-90e5-3f1a-09ed8d4aea27" [ 1015.912986] env[62385]: _type = "Task" [ 1015.912986] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.921632] env[62385]: DEBUG oslo_vmware.api [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c19ca8-f566-90e5-3f1a-09ed8d4aea27, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.165520] env[62385]: DEBUG nova.scheduler.client.report [None req-c637a0de-3a26-483f-8836-57b0fe6d6848 tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1016.327228] env[62385]: DEBUG oslo_concurrency.lockutils [req-0ca9725b-3fbd-48ae-b282-b6f57fb258a5 req-7596b8b9-e5d3-41bc-bffd-230c62d0c891 service nova] Releasing lock "refresh_cache-b0a6a3b9-05e4-4190-9cc6-4237dee07f91" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.327639] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206336, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.517124} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.327883] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 5daf0a99-0c2a-40d8-afc7-1998e21e32d8/5daf0a99-0c2a-40d8-afc7-1998e21e32d8.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1016.328133] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1016.328408] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-47119077-8219-4740-8090-f9ff4ed873b8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.336074] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1016.336074] env[62385]: value = "task-1206337" [ 1016.336074] env[62385]: _type = "Task" [ 1016.336074] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.345158] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206337, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.423588] env[62385]: DEBUG oslo_vmware.api [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c19ca8-f566-90e5-3f1a-09ed8d4aea27, 'name': SearchDatastore_Task, 'duration_secs': 0.010777} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.424424] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4caa9efa-9a4c-4d63-853e-9de81cda7bae {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.430692] env[62385]: DEBUG oslo_vmware.api [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 1016.430692] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521d3d4a-ad5f-1d8d-456f-77d87f2cc8bc" [ 1016.430692] env[62385]: _type = "Task" [ 1016.430692] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.439817] env[62385]: DEBUG oslo_vmware.api [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521d3d4a-ad5f-1d8d-456f-77d87f2cc8bc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.670906] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c637a0de-3a26-483f-8836-57b0fe6d6848 tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.779s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.674048] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.223s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.675146] env[62385]: INFO nova.compute.claims [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1016.693028] env[62385]: INFO nova.scheduler.client.report [None req-c637a0de-3a26-483f-8836-57b0fe6d6848 tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Deleted allocations for instance 784589bb-a668-4674-83b8-6219def15067 [ 1016.845664] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206337, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.0752} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.845935] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1016.846715] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17f17279-3f01-4b0b-ad6b-ff53e964fa2e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.868606] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] 5daf0a99-0c2a-40d8-afc7-1998e21e32d8/5daf0a99-0c2a-40d8-afc7-1998e21e32d8.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1016.869200] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9ea055b6-2c24-40b8-a06f-6b865544dbdd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.889924] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1016.889924] env[62385]: value = "task-1206338" [ 1016.889924] env[62385]: _type = "Task" [ 1016.889924] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.899323] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206338, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.912105] env[62385]: DEBUG nova.compute.manager [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1016.943331] env[62385]: DEBUG nova.virt.hardware [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1016.943586] env[62385]: DEBUG nova.virt.hardware [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1016.943747] env[62385]: DEBUG nova.virt.hardware [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1016.943937] env[62385]: DEBUG nova.virt.hardware [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1016.944108] env[62385]: DEBUG nova.virt.hardware [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1016.944266] env[62385]: DEBUG nova.virt.hardware [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1016.944497] env[62385]: DEBUG nova.virt.hardware [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1016.944640] env[62385]: DEBUG nova.virt.hardware [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1016.944811] env[62385]: DEBUG nova.virt.hardware [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1016.945065] env[62385]: DEBUG nova.virt.hardware [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1016.945170] env[62385]: DEBUG nova.virt.hardware [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1016.946022] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b2f182c-8e77-44a9-9d45-0d49441bdca3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.952300] env[62385]: DEBUG oslo_vmware.api [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521d3d4a-ad5f-1d8d-456f-77d87f2cc8bc, 'name': SearchDatastore_Task, 'duration_secs': 0.010609} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.952941] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.953215] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] b0a6a3b9-05e4-4190-9cc6-4237dee07f91/b0a6a3b9-05e4-4190-9cc6-4237dee07f91.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1016.953490] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-84632c61-3d71-4ea5-ad19-9af97010e9c9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.959028] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e415b05-dbf5-4317-8e7c-8f232b84d091 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.964214] env[62385]: DEBUG oslo_vmware.api [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 1016.964214] env[62385]: value = "task-1206339" [ 1016.964214] env[62385]: _type = "Task" [ 1016.964214] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.980387] env[62385]: DEBUG oslo_vmware.api [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206339, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.137019] env[62385]: DEBUG oslo_concurrency.lockutils [None req-57468fc3-c541-446a-91b3-64d71c7db5a6 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Acquiring lock "a1c71d8b-0953-4cc2-961d-0c2b2fc11110" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.137380] env[62385]: DEBUG oslo_concurrency.lockutils [None req-57468fc3-c541-446a-91b3-64d71c7db5a6 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "a1c71d8b-0953-4cc2-961d-0c2b2fc11110" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.202990] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c637a0de-3a26-483f-8836-57b0fe6d6848 tempest-ServerMetadataNegativeTestJSON-577309250 tempest-ServerMetadataNegativeTestJSON-577309250-project-member] Lock "784589bb-a668-4674-83b8-6219def15067" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.274s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.404855] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206338, 'name': ReconfigVM_Task, 'duration_secs': 0.312424} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.405254] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Reconfigured VM instance instance-0000005d to attach disk [datastore1] 5daf0a99-0c2a-40d8-afc7-1998e21e32d8/5daf0a99-0c2a-40d8-afc7-1998e21e32d8.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1017.406886] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'size': 0, 'encrypted': False, 'disk_bus': None, 'encryption_secret_uuid': None, 'boot_index': 0, 'encryption_options': None, 'device_type': 'disk', 'guest_format': None, 'device_name': '/dev/sda', 'encryption_format': None, 'image_id': 'c0abbb8d-77e7-4f22-a256-0faf0a781109'}], 'ephemerals': [], 'block_device_mapping': [{'disk_bus': None, 'attachment_id': '4cfd7da2-3a7e-4eae-a550-c47d2cb64a37', 'boot_index': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261287', 'volume_id': '9621f12c-980a-4762-8ea2-c4fff11dc20e', 'name': 'volume-9621f12c-980a-4762-8ea2-c4fff11dc20e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5daf0a99-0c2a-40d8-afc7-1998e21e32d8', 'attached_at': '', 'detached_at': '', 'volume_id': '9621f12c-980a-4762-8ea2-c4fff11dc20e', 'serial': '9621f12c-980a-4762-8ea2-c4fff11dc20e'}, 'delete_on_termination': False, 'mount_device': '/dev/sdb', 'device_type': None, 'guest_format': None, 'volume_type': None}], 'swap': None} {{(pid=62385) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1017.407139] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Volume attach. Driver type: vmdk {{(pid=62385) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1017.407355] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261287', 'volume_id': '9621f12c-980a-4762-8ea2-c4fff11dc20e', 'name': 'volume-9621f12c-980a-4762-8ea2-c4fff11dc20e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5daf0a99-0c2a-40d8-afc7-1998e21e32d8', 'attached_at': '', 'detached_at': '', 'volume_id': '9621f12c-980a-4762-8ea2-c4fff11dc20e', 'serial': '9621f12c-980a-4762-8ea2-c4fff11dc20e'} {{(pid=62385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1017.408309] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c1220cd-e5a9-4038-b91d-a9f3ebc99e14 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.427699] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e4d06b4-e090-47c8-b22b-b5682b198e14 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.455463] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] volume-9621f12c-980a-4762-8ea2-c4fff11dc20e/volume-9621f12c-980a-4762-8ea2-c4fff11dc20e.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1017.456016] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-54bfc52a-ad72-4269-9d11-166d4792d31a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.482259] env[62385]: DEBUG oslo_vmware.api [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206339, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.507504} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.483601] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] b0a6a3b9-05e4-4190-9cc6-4237dee07f91/b0a6a3b9-05e4-4190-9cc6-4237dee07f91.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1017.483844] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1017.484188] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1017.484188] env[62385]: value = "task-1206341" [ 1017.484188] env[62385]: _type = "Task" [ 1017.484188] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.484398] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ce8771c7-e791-4f5b-8b5d-4a9284f9290b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.494904] env[62385]: DEBUG oslo_vmware.api [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 1017.494904] env[62385]: value = "task-1206342" [ 1017.494904] env[62385]: _type = "Task" [ 1017.494904] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.498242] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206341, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.509274] env[62385]: DEBUG oslo_vmware.api [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206342, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.640859] env[62385]: DEBUG nova.compute.utils [None req-57468fc3-c541-446a-91b3-64d71c7db5a6 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1017.721671] env[62385]: DEBUG nova.compute.manager [req-954be31f-4be2-4b4b-9098-55bb894ca4c7 req-c28bcc0f-9687-43be-8351-834178b32997 service nova] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Received event network-vif-plugged-c308701b-9b9e-4ff3-9a01-c27aabf54f0d {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1017.721888] env[62385]: DEBUG oslo_concurrency.lockutils [req-954be31f-4be2-4b4b-9098-55bb894ca4c7 req-c28bcc0f-9687-43be-8351-834178b32997 service nova] Acquiring lock "68748d25-017f-4b58-9984-70264abbc5ba-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.722595] env[62385]: DEBUG oslo_concurrency.lockutils [req-954be31f-4be2-4b4b-9098-55bb894ca4c7 req-c28bcc0f-9687-43be-8351-834178b32997 service nova] Lock "68748d25-017f-4b58-9984-70264abbc5ba-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.722595] env[62385]: DEBUG oslo_concurrency.lockutils [req-954be31f-4be2-4b4b-9098-55bb894ca4c7 req-c28bcc0f-9687-43be-8351-834178b32997 service nova] Lock "68748d25-017f-4b58-9984-70264abbc5ba-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.722595] env[62385]: DEBUG nova.compute.manager [req-954be31f-4be2-4b4b-9098-55bb894ca4c7 req-c28bcc0f-9687-43be-8351-834178b32997 service nova] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] No waiting events found dispatching network-vif-plugged-c308701b-9b9e-4ff3-9a01-c27aabf54f0d {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1017.722595] env[62385]: WARNING nova.compute.manager [req-954be31f-4be2-4b4b-9098-55bb894ca4c7 req-c28bcc0f-9687-43be-8351-834178b32997 service nova] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Received unexpected event network-vif-plugged-c308701b-9b9e-4ff3-9a01-c27aabf54f0d for instance with vm_state building and task_state spawning. [ 1017.793465] env[62385]: DEBUG nova.network.neutron [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Successfully updated port: c308701b-9b9e-4ff3-9a01-c27aabf54f0d {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1017.890248] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd5973dc-e910-4934-a2a8-d04e2c648689 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.899026] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4232f12-19ac-4512-9ea8-829c0ed4fe4a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.930627] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3b85fb4-3919-491b-bfc9-c3f52eefb174 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.938298] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-593950e7-985a-4c34-976d-b70d84bcbf7c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.951987] env[62385]: DEBUG nova.compute.provider_tree [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1017.996043] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206341, 'name': ReconfigVM_Task, 'duration_secs': 0.355351} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.996395] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Reconfigured VM instance instance-0000005d to attach disk [datastore1] volume-9621f12c-980a-4762-8ea2-c4fff11dc20e/volume-9621f12c-980a-4762-8ea2-c4fff11dc20e.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1018.001202] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-62d767c0-e730-4dac-9f28-f71b2d41d9d5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.019724] env[62385]: DEBUG oslo_vmware.api [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206342, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061246} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.020874] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1018.021272] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1018.021272] env[62385]: value = "task-1206343" [ 1018.021272] env[62385]: _type = "Task" [ 1018.021272] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.021963] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc23f0a9-7244-499a-b56b-723c9a20a916 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.032545] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206343, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.051892] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] b0a6a3b9-05e4-4190-9cc6-4237dee07f91/b0a6a3b9-05e4-4190-9cc6-4237dee07f91.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1018.051892] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cb374875-ba0e-46a1-9450-238744b1af97 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.072965] env[62385]: DEBUG oslo_vmware.api [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 1018.072965] env[62385]: value = "task-1206344" [ 1018.072965] env[62385]: _type = "Task" [ 1018.072965] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.081603] env[62385]: DEBUG oslo_vmware.api [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206344, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.144495] env[62385]: DEBUG oslo_concurrency.lockutils [None req-57468fc3-c541-446a-91b3-64d71c7db5a6 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "a1c71d8b-0953-4cc2-961d-0c2b2fc11110" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.296386] env[62385]: DEBUG oslo_concurrency.lockutils [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "refresh_cache-68748d25-017f-4b58-9984-70264abbc5ba" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.296760] env[62385]: DEBUG oslo_concurrency.lockutils [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquired lock "refresh_cache-68748d25-017f-4b58-9984-70264abbc5ba" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.297039] env[62385]: DEBUG nova.network.neutron [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1018.455045] env[62385]: DEBUG nova.scheduler.client.report [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1018.535650] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206343, 'name': ReconfigVM_Task, 'duration_secs': 0.177594} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.535650] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261287', 'volume_id': '9621f12c-980a-4762-8ea2-c4fff11dc20e', 'name': 'volume-9621f12c-980a-4762-8ea2-c4fff11dc20e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5daf0a99-0c2a-40d8-afc7-1998e21e32d8', 'attached_at': '', 'detached_at': '', 'volume_id': '9621f12c-980a-4762-8ea2-c4fff11dc20e', 'serial': '9621f12c-980a-4762-8ea2-c4fff11dc20e'} {{(pid=62385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1018.535650] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-99c8b512-b7de-4212-bb14-4bf0579d8b18 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.543973] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1018.543973] env[62385]: value = "task-1206345" [ 1018.543973] env[62385]: _type = "Task" [ 1018.543973] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.555162] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206345, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.583025] env[62385]: DEBUG oslo_vmware.api [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206344, 'name': ReconfigVM_Task, 'duration_secs': 0.494433} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.583025] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Reconfigured VM instance instance-00000067 to attach disk [datastore1] b0a6a3b9-05e4-4190-9cc6-4237dee07f91/b0a6a3b9-05e4-4190-9cc6-4237dee07f91.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1018.583345] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-198bddfc-ac78-4144-afb1-d7f4667f4b6e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.589928] env[62385]: DEBUG oslo_vmware.api [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 1018.589928] env[62385]: value = "task-1206346" [ 1018.589928] env[62385]: _type = "Task" [ 1018.589928] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.599559] env[62385]: DEBUG oslo_vmware.api [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206346, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.829285] env[62385]: DEBUG nova.network.neutron [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1018.961877] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.288s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.962460] env[62385]: DEBUG nova.compute.manager [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1018.965120] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 8.464s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.965297] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.969078] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62385) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1018.969078] env[62385]: DEBUG oslo_concurrency.lockutils [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.582s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.969078] env[62385]: DEBUG oslo_concurrency.lockutils [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.969078] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7525a56b-c44c-4f42-aefb-21ae42b9f33a tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 5.336s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.973620] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-330d7fc8-a1c1-4eb2-9939-953a049e9a44 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.986188] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6c8fc24-fc5e-497b-9b53-4e0da7357161 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.998262] env[62385]: INFO nova.scheduler.client.report [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Deleted allocations for instance b1e0baf3-643c-49c2-8a80-8de07d8527e7 [ 1019.009837] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98e9b889-fd43-44f5-abe2-c59f7e2556e7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.018227] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78ab6979-88ee-4b7b-9297-f6e0bb6abf6c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.050077] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180385MB free_disk=155GB free_vcpus=48 pci_devices=None {{(pid=62385) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1019.050293] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.051356] env[62385]: DEBUG nova.network.neutron [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Updating instance_info_cache with network_info: [{"id": "c308701b-9b9e-4ff3-9a01-c27aabf54f0d", "address": "fa:16:3e:9a:ed:8c", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc308701b-9b", "ovs_interfaceid": "c308701b-9b9e-4ff3-9a01-c27aabf54f0d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.061318] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206345, 'name': Rename_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.100713] env[62385]: DEBUG oslo_vmware.api [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206346, 'name': Rename_Task, 'duration_secs': 0.144336} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.101011] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1019.101277] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-da0c2aab-d764-4c29-81ed-6900435c6793 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.108060] env[62385]: DEBUG oslo_vmware.api [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 1019.108060] env[62385]: value = "task-1206347" [ 1019.108060] env[62385]: _type = "Task" [ 1019.108060] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.115739] env[62385]: DEBUG oslo_vmware.api [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206347, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.211910] env[62385]: DEBUG oslo_concurrency.lockutils [None req-57468fc3-c541-446a-91b3-64d71c7db5a6 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Acquiring lock "a1c71d8b-0953-4cc2-961d-0c2b2fc11110" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.212253] env[62385]: DEBUG oslo_concurrency.lockutils [None req-57468fc3-c541-446a-91b3-64d71c7db5a6 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "a1c71d8b-0953-4cc2-961d-0c2b2fc11110" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.212680] env[62385]: INFO nova.compute.manager [None req-57468fc3-c541-446a-91b3-64d71c7db5a6 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Attaching volume 3b02b9da-5324-4812-abd7-09ff15288694 to /dev/sdb [ 1019.252311] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7107e17-3cce-46a9-9c52-e8d700500779 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.259786] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef4223b3-4b6b-4075-a701-733f8fe15c3a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.274636] env[62385]: DEBUG nova.virt.block_device [None req-57468fc3-c541-446a-91b3-64d71c7db5a6 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Updating existing volume attachment record: 59f332e2-dddd-4e2d-8402-abd9a33d6658 {{(pid=62385) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1019.472084] env[62385]: DEBUG nova.compute.utils [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1019.473494] env[62385]: DEBUG nova.compute.manager [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1019.473667] env[62385]: DEBUG nova.network.neutron [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1019.516690] env[62385]: DEBUG oslo_concurrency.lockutils [None req-95a46163-9ba3-4eee-a231-423f8565b12d tempest-DeleteServersTestJSON-45311166 tempest-DeleteServersTestJSON-45311166-project-member] Lock "b1e0baf3-643c-49c2-8a80-8de07d8527e7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.052s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.557438] env[62385]: DEBUG oslo_concurrency.lockutils [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Releasing lock "refresh_cache-68748d25-017f-4b58-9984-70264abbc5ba" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.557868] env[62385]: DEBUG nova.compute.manager [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Instance network_info: |[{"id": "c308701b-9b9e-4ff3-9a01-c27aabf54f0d", "address": "fa:16:3e:9a:ed:8c", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc308701b-9b", "ovs_interfaceid": "c308701b-9b9e-4ff3-9a01-c27aabf54f0d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1019.562510] env[62385]: DEBUG nova.policy [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5e83312d226945c99d05fcc7f74c0978', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bf8e3c942d3445919cfbe988cca84e90', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 1019.564292] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9a:ed:8c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '685b4083-b748-41fb-a68a-273b1073fa28', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c308701b-9b9e-4ff3-9a01-c27aabf54f0d', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1019.572180] env[62385]: DEBUG oslo.service.loopingcall [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1019.572482] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206345, 'name': Rename_Task, 'duration_secs': 0.757631} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.573056] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1019.573380] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1019.573951] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8d8ea411-1d15-45d2-9fc0-184f28d561c3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.588408] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c359ad24-0d3f-46da-8201-aff7aa3dfe69 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.600058] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1019.600058] env[62385]: value = "task-1206350" [ 1019.600058] env[62385]: _type = "Task" [ 1019.600058] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.600058] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1019.600058] env[62385]: value = "task-1206349" [ 1019.600058] env[62385]: _type = "Task" [ 1019.600058] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.618450] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206349, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.622830] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206350, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.630239] env[62385]: DEBUG oslo_vmware.api [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206347, 'name': PowerOnVM_Task, 'duration_secs': 0.517215} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.630545] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1019.630752] env[62385]: INFO nova.compute.manager [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Took 7.27 seconds to spawn the instance on the hypervisor. [ 1019.630958] env[62385]: DEBUG nova.compute.manager [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1019.631784] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c784e12e-ab90-4cb7-b498-5a2ccb6bbcc5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.686554] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-604e6b77-5359-41b7-bb1d-f62a9644acac {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.696590] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-009544cb-29d5-40b8-8299-86d5c70a811f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.733412] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-540d9590-35b2-4256-aa04-b6110cb2702e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.743893] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42319848-76a2-4c71-a15f-cccd81b5c6a3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.761851] env[62385]: DEBUG nova.compute.provider_tree [None req-7525a56b-c44c-4f42-aefb-21ae42b9f33a tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1019.765364] env[62385]: DEBUG nova.compute.manager [req-900707f2-20f1-4a04-a4b0-666760e0792b req-65230848-acf1-4c77-93e5-4ce3a2833351 service nova] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Received event network-changed-c308701b-9b9e-4ff3-9a01-c27aabf54f0d {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1019.767039] env[62385]: DEBUG nova.compute.manager [req-900707f2-20f1-4a04-a4b0-666760e0792b req-65230848-acf1-4c77-93e5-4ce3a2833351 service nova] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Refreshing instance network info cache due to event network-changed-c308701b-9b9e-4ff3-9a01-c27aabf54f0d. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1019.767039] env[62385]: DEBUG oslo_concurrency.lockutils [req-900707f2-20f1-4a04-a4b0-666760e0792b req-65230848-acf1-4c77-93e5-4ce3a2833351 service nova] Acquiring lock "refresh_cache-68748d25-017f-4b58-9984-70264abbc5ba" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.767039] env[62385]: DEBUG oslo_concurrency.lockutils [req-900707f2-20f1-4a04-a4b0-666760e0792b req-65230848-acf1-4c77-93e5-4ce3a2833351 service nova] Acquired lock "refresh_cache-68748d25-017f-4b58-9984-70264abbc5ba" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.767039] env[62385]: DEBUG nova.network.neutron [req-900707f2-20f1-4a04-a4b0-666760e0792b req-65230848-acf1-4c77-93e5-4ce3a2833351 service nova] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Refreshing network info cache for port c308701b-9b9e-4ff3-9a01-c27aabf54f0d {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1019.980064] env[62385]: DEBUG nova.compute.manager [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1020.012756] env[62385]: DEBUG nova.network.neutron [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Successfully created port: 4af1e8e6-edeb-47ff-b1b1-f4f974a69544 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1020.118161] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206350, 'name': CreateVM_Task, 'duration_secs': 0.370495} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.121250] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1020.121565] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206349, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.123018] env[62385]: DEBUG oslo_concurrency.lockutils [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.123018] env[62385]: DEBUG oslo_concurrency.lockutils [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.123018] env[62385]: DEBUG oslo_concurrency.lockutils [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1020.123018] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db31b23e-6330-4e6a-94f2-b808d86e7683 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.128324] env[62385]: DEBUG oslo_vmware.api [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 1020.128324] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52cd87aa-0f56-78d1-b86a-6cd3101a1084" [ 1020.128324] env[62385]: _type = "Task" [ 1020.128324] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.137411] env[62385]: DEBUG oslo_vmware.api [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52cd87aa-0f56-78d1-b86a-6cd3101a1084, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.158545] env[62385]: INFO nova.compute.manager [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Took 14.60 seconds to build instance. [ 1020.269380] env[62385]: DEBUG nova.scheduler.client.report [None req-7525a56b-c44c-4f42-aefb-21ae42b9f33a tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1020.392746] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.491109] env[62385]: DEBUG nova.network.neutron [req-900707f2-20f1-4a04-a4b0-666760e0792b req-65230848-acf1-4c77-93e5-4ce3a2833351 service nova] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Updated VIF entry in instance network info cache for port c308701b-9b9e-4ff3-9a01-c27aabf54f0d. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1020.491109] env[62385]: DEBUG nova.network.neutron [req-900707f2-20f1-4a04-a4b0-666760e0792b req-65230848-acf1-4c77-93e5-4ce3a2833351 service nova] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Updating instance_info_cache with network_info: [{"id": "c308701b-9b9e-4ff3-9a01-c27aabf54f0d", "address": "fa:16:3e:9a:ed:8c", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc308701b-9b", "ovs_interfaceid": "c308701b-9b9e-4ff3-9a01-c27aabf54f0d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.612104] env[62385]: DEBUG oslo_vmware.api [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206349, 'name': PowerOnVM_Task, 'duration_secs': 0.651505} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.612409] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1020.612633] env[62385]: DEBUG nova.compute.manager [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1020.613443] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da0535c0-a2a9-4487-a9fb-4975c6f9e41d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.639351] env[62385]: DEBUG oslo_vmware.api [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52cd87aa-0f56-78d1-b86a-6cd3101a1084, 'name': SearchDatastore_Task, 'duration_secs': 0.015948} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.639665] env[62385]: DEBUG oslo_concurrency.lockutils [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.639900] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1020.643606] env[62385]: DEBUG oslo_concurrency.lockutils [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.643606] env[62385]: DEBUG oslo_concurrency.lockutils [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.643606] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1020.643606] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dcfa33ea-434f-4709-ac4b-3f69233e991e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.650673] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1020.650873] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1020.651700] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-42bf2d3b-5beb-4914-a929-d917c03e85d7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.658303] env[62385]: DEBUG oslo_vmware.api [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 1020.658303] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]522b09dd-7960-db6d-7dbc-e187a2131c03" [ 1020.658303] env[62385]: _type = "Task" [ 1020.658303] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.661923] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2ff0dcfd-8d18-490b-b0e2-b7cc8fba49db tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "b0a6a3b9-05e4-4190-9cc6-4237dee07f91" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.114s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.667337] env[62385]: DEBUG oslo_vmware.api [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]522b09dd-7960-db6d-7dbc-e187a2131c03, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.990369] env[62385]: DEBUG nova.compute.manager [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1020.992988] env[62385]: DEBUG oslo_concurrency.lockutils [req-900707f2-20f1-4a04-a4b0-666760e0792b req-65230848-acf1-4c77-93e5-4ce3a2833351 service nova] Releasing lock "refresh_cache-68748d25-017f-4b58-9984-70264abbc5ba" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.020610] env[62385]: DEBUG nova.virt.hardware [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1021.021082] env[62385]: DEBUG nova.virt.hardware [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1021.021386] env[62385]: DEBUG nova.virt.hardware [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1021.021686] env[62385]: DEBUG nova.virt.hardware [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1021.021943] env[62385]: DEBUG nova.virt.hardware [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1021.022234] env[62385]: DEBUG nova.virt.hardware [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1021.022600] env[62385]: DEBUG nova.virt.hardware [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1021.022874] env[62385]: DEBUG nova.virt.hardware [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1021.025198] env[62385]: DEBUG nova.virt.hardware [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1021.025198] env[62385]: DEBUG nova.virt.hardware [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1021.025198] env[62385]: DEBUG nova.virt.hardware [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1021.025198] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a9c4b70-1c3b-4f8f-9322-3ee378beb366 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.033616] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-986dffe5-0d59-41f4-a65e-87260f31a9db {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.132614] env[62385]: DEBUG oslo_concurrency.lockutils [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.170009] env[62385]: DEBUG oslo_vmware.api [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]522b09dd-7960-db6d-7dbc-e187a2131c03, 'name': SearchDatastore_Task, 'duration_secs': 0.019097} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.170933] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b13ca9f-fd02-4787-a432-009721e1a741 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.178112] env[62385]: DEBUG oslo_vmware.api [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 1021.178112] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e8e617-9210-2aa1-3037-081978a5b220" [ 1021.178112] env[62385]: _type = "Task" [ 1021.178112] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.187073] env[62385]: DEBUG oslo_vmware.api [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e8e617-9210-2aa1-3037-081978a5b220, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.282598] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7525a56b-c44c-4f42-aefb-21ae42b9f33a tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.315s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.288334] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.589s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.289713] env[62385]: INFO nova.compute.claims [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1021.382890] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c6808a79-c5d5-48a7-9431-20da78d448b1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "b0a6a3b9-05e4-4190-9cc6-4237dee07f91" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.382890] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c6808a79-c5d5-48a7-9431-20da78d448b1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "b0a6a3b9-05e4-4190-9cc6-4237dee07f91" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.383207] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c6808a79-c5d5-48a7-9431-20da78d448b1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "b0a6a3b9-05e4-4190-9cc6-4237dee07f91-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.383281] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c6808a79-c5d5-48a7-9431-20da78d448b1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "b0a6a3b9-05e4-4190-9cc6-4237dee07f91-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.384033] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c6808a79-c5d5-48a7-9431-20da78d448b1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "b0a6a3b9-05e4-4190-9cc6-4237dee07f91-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.386206] env[62385]: INFO nova.compute.manager [None req-c6808a79-c5d5-48a7-9431-20da78d448b1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Terminating instance [ 1021.387931] env[62385]: DEBUG nova.compute.manager [None req-c6808a79-c5d5-48a7-9431-20da78d448b1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1021.388153] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c6808a79-c5d5-48a7-9431-20da78d448b1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1021.388968] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a81de805-de14-44a3-9b25-8a2ba53dea7b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.397347] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6808a79-c5d5-48a7-9431-20da78d448b1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1021.397594] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3f7221b4-0db3-4148-b23c-8469fce06e87 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.405917] env[62385]: DEBUG oslo_vmware.api [None req-c6808a79-c5d5-48a7-9431-20da78d448b1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 1021.405917] env[62385]: value = "task-1206351" [ 1021.405917] env[62385]: _type = "Task" [ 1021.405917] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.414453] env[62385]: DEBUG oslo_vmware.api [None req-c6808a79-c5d5-48a7-9431-20da78d448b1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206351, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.691125] env[62385]: DEBUG oslo_vmware.api [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e8e617-9210-2aa1-3037-081978a5b220, 'name': SearchDatastore_Task, 'duration_secs': 0.010754} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.691522] env[62385]: DEBUG oslo_concurrency.lockutils [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.691717] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 68748d25-017f-4b58-9984-70264abbc5ba/68748d25-017f-4b58-9984-70264abbc5ba.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1021.692056] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6a4063f9-4098-438c-aa0c-c15f5de9d3e9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.700108] env[62385]: DEBUG oslo_vmware.api [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 1021.700108] env[62385]: value = "task-1206353" [ 1021.700108] env[62385]: _type = "Task" [ 1021.700108] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.713384] env[62385]: DEBUG oslo_vmware.api [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206353, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.850698] env[62385]: INFO nova.scheduler.client.report [None req-7525a56b-c44c-4f42-aefb-21ae42b9f33a tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Deleted allocation for migration a29e1d61-503d-4be1-9d07-d708bf2b943e [ 1021.919048] env[62385]: DEBUG oslo_vmware.api [None req-c6808a79-c5d5-48a7-9431-20da78d448b1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206351, 'name': PowerOffVM_Task, 'duration_secs': 0.163864} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.919652] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6808a79-c5d5-48a7-9431-20da78d448b1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1021.919900] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c6808a79-c5d5-48a7-9431-20da78d448b1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1021.920354] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3e5de41a-4d4e-4448-99a7-72a799833398 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.977322] env[62385]: DEBUG nova.compute.manager [req-f5a42cfb-00c7-4d0a-9dfd-0d778f4cb8ce req-0a8839ca-8fff-4bd6-af6e-a19cbb5ce3d0 service nova] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Received event network-vif-plugged-4af1e8e6-edeb-47ff-b1b1-f4f974a69544 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1021.977571] env[62385]: DEBUG oslo_concurrency.lockutils [req-f5a42cfb-00c7-4d0a-9dfd-0d778f4cb8ce req-0a8839ca-8fff-4bd6-af6e-a19cbb5ce3d0 service nova] Acquiring lock "47311a9c-74d2-400d-a22d-9ef55b14edb5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.977747] env[62385]: DEBUG oslo_concurrency.lockutils [req-f5a42cfb-00c7-4d0a-9dfd-0d778f4cb8ce req-0a8839ca-8fff-4bd6-af6e-a19cbb5ce3d0 service nova] Lock "47311a9c-74d2-400d-a22d-9ef55b14edb5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.978696] env[62385]: DEBUG oslo_concurrency.lockutils [req-f5a42cfb-00c7-4d0a-9dfd-0d778f4cb8ce req-0a8839ca-8fff-4bd6-af6e-a19cbb5ce3d0 service nova] Lock "47311a9c-74d2-400d-a22d-9ef55b14edb5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.978696] env[62385]: DEBUG nova.compute.manager [req-f5a42cfb-00c7-4d0a-9dfd-0d778f4cb8ce req-0a8839ca-8fff-4bd6-af6e-a19cbb5ce3d0 service nova] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] No waiting events found dispatching network-vif-plugged-4af1e8e6-edeb-47ff-b1b1-f4f974a69544 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1021.978696] env[62385]: WARNING nova.compute.manager [req-f5a42cfb-00c7-4d0a-9dfd-0d778f4cb8ce req-0a8839ca-8fff-4bd6-af6e-a19cbb5ce3d0 service nova] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Received unexpected event network-vif-plugged-4af1e8e6-edeb-47ff-b1b1-f4f974a69544 for instance with vm_state building and task_state spawning. [ 1022.023348] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c6808a79-c5d5-48a7-9431-20da78d448b1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1022.023594] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c6808a79-c5d5-48a7-9431-20da78d448b1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1022.023787] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6808a79-c5d5-48a7-9431-20da78d448b1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Deleting the datastore file [datastore1] b0a6a3b9-05e4-4190-9cc6-4237dee07f91 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1022.024187] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-44d07e1c-7ccb-4447-8eac-a71ac7d84add {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.034723] env[62385]: DEBUG oslo_vmware.api [None req-c6808a79-c5d5-48a7-9431-20da78d448b1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for the task: (returnval){ [ 1022.034723] env[62385]: value = "task-1206355" [ 1022.034723] env[62385]: _type = "Task" [ 1022.034723] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.044351] env[62385]: DEBUG oslo_vmware.api [None req-c6808a79-c5d5-48a7-9431-20da78d448b1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206355, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.215548] env[62385]: DEBUG oslo_vmware.api [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206353, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.473969} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.218101] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 68748d25-017f-4b58-9984-70264abbc5ba/68748d25-017f-4b58-9984-70264abbc5ba.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1022.218301] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1022.218601] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-53a33af5-5f1d-4fd1-beaf-68d2b5050f4b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.227131] env[62385]: DEBUG oslo_vmware.api [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 1022.227131] env[62385]: value = "task-1206356" [ 1022.227131] env[62385]: _type = "Task" [ 1022.227131] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.241729] env[62385]: DEBUG oslo_vmware.api [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206356, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.362284] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7525a56b-c44c-4f42-aefb-21ae42b9f33a tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "9b9fd167-d843-4b9d-9e4f-dbde5d3d629e" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 11.612s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.496340] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ce0d470-b33d-4040-97f1-82a6e3dc3321 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.505530] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e1bad4a-26f9-4f65-b567-bbc7ff6c57ee {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.543429] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d96526d4-5e4e-4d90-b18c-37bad849a2c4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.551791] env[62385]: DEBUG oslo_vmware.api [None req-c6808a79-c5d5-48a7-9431-20da78d448b1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Task: {'id': task-1206355, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.256753} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.557460] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6808a79-c5d5-48a7-9431-20da78d448b1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1022.557534] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c6808a79-c5d5-48a7-9431-20da78d448b1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1022.557717] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c6808a79-c5d5-48a7-9431-20da78d448b1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1022.558384] env[62385]: INFO nova.compute.manager [None req-c6808a79-c5d5-48a7-9431-20da78d448b1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1022.558384] env[62385]: DEBUG oslo.service.loopingcall [None req-c6808a79-c5d5-48a7-9431-20da78d448b1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1022.558969] env[62385]: DEBUG nova.network.neutron [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Successfully updated port: 4af1e8e6-edeb-47ff-b1b1-f4f974a69544 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1022.560187] env[62385]: DEBUG nova.compute.manager [-] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1022.560290] env[62385]: DEBUG nova.network.neutron [-] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1022.569064] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26926de9-051c-4b0d-8183-6666c3481965 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.584913] env[62385]: DEBUG nova.compute.provider_tree [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1022.588512] env[62385]: DEBUG nova.compute.manager [req-a9116891-734d-4168-b799-f8187fb5d0f2 req-fe728fd6-f7e8-48aa-a792-d25fe2412259 service nova] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Received event network-changed-4af1e8e6-edeb-47ff-b1b1-f4f974a69544 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1022.588706] env[62385]: DEBUG nova.compute.manager [req-a9116891-734d-4168-b799-f8187fb5d0f2 req-fe728fd6-f7e8-48aa-a792-d25fe2412259 service nova] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Refreshing instance network info cache due to event network-changed-4af1e8e6-edeb-47ff-b1b1-f4f974a69544. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1022.588917] env[62385]: DEBUG oslo_concurrency.lockutils [req-a9116891-734d-4168-b799-f8187fb5d0f2 req-fe728fd6-f7e8-48aa-a792-d25fe2412259 service nova] Acquiring lock "refresh_cache-47311a9c-74d2-400d-a22d-9ef55b14edb5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1022.589074] env[62385]: DEBUG oslo_concurrency.lockutils [req-a9116891-734d-4168-b799-f8187fb5d0f2 req-fe728fd6-f7e8-48aa-a792-d25fe2412259 service nova] Acquired lock "refresh_cache-47311a9c-74d2-400d-a22d-9ef55b14edb5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.589241] env[62385]: DEBUG nova.network.neutron [req-a9116891-734d-4168-b799-f8187fb5d0f2 req-fe728fd6-f7e8-48aa-a792-d25fe2412259 service nova] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Refreshing network info cache for port 4af1e8e6-edeb-47ff-b1b1-f4f974a69544 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1022.737874] env[62385]: DEBUG oslo_vmware.api [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206356, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.093512} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.738752] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1022.740794] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c324f577-be06-43b5-89da-0313ae7be437 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.766123] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] 68748d25-017f-4b58-9984-70264abbc5ba/68748d25-017f-4b58-9984-70264abbc5ba.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1022.766614] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-94fc666c-0554-444f-aecc-242e969a73a7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.792674] env[62385]: DEBUG oslo_vmware.api [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 1022.792674] env[62385]: value = "task-1206357" [ 1022.792674] env[62385]: _type = "Task" [ 1022.792674] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.801392] env[62385]: DEBUG oslo_vmware.api [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206357, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.859896] env[62385]: DEBUG oslo_concurrency.lockutils [None req-db9034be-4bf6-4cb8-a612-9f11df348d57 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "5daf0a99-0c2a-40d8-afc7-1998e21e32d8" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.860368] env[62385]: DEBUG oslo_concurrency.lockutils [None req-db9034be-4bf6-4cb8-a612-9f11df348d57 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "5daf0a99-0c2a-40d8-afc7-1998e21e32d8" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.071733] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "refresh_cache-47311a9c-74d2-400d-a22d-9ef55b14edb5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1023.091971] env[62385]: DEBUG nova.scheduler.client.report [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1023.124283] env[62385]: DEBUG nova.network.neutron [req-a9116891-734d-4168-b799-f8187fb5d0f2 req-fe728fd6-f7e8-48aa-a792-d25fe2412259 service nova] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1023.199644] env[62385]: DEBUG nova.network.neutron [req-a9116891-734d-4168-b799-f8187fb5d0f2 req-fe728fd6-f7e8-48aa-a792-d25fe2412259 service nova] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.228188] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c02c14cc-903b-4279-9294-6fa98f845aa2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "9b9fd167-d843-4b9d-9e4f-dbde5d3d629e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.228561] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c02c14cc-903b-4279-9294-6fa98f845aa2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "9b9fd167-d843-4b9d-9e4f-dbde5d3d629e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.228833] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c02c14cc-903b-4279-9294-6fa98f845aa2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "9b9fd167-d843-4b9d-9e4f-dbde5d3d629e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.229061] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c02c14cc-903b-4279-9294-6fa98f845aa2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "9b9fd167-d843-4b9d-9e4f-dbde5d3d629e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.229252] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c02c14cc-903b-4279-9294-6fa98f845aa2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "9b9fd167-d843-4b9d-9e4f-dbde5d3d629e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.235681] env[62385]: INFO nova.compute.manager [None req-c02c14cc-903b-4279-9294-6fa98f845aa2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Terminating instance [ 1023.237765] env[62385]: DEBUG nova.compute.manager [None req-c02c14cc-903b-4279-9294-6fa98f845aa2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1023.238033] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c02c14cc-903b-4279-9294-6fa98f845aa2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1023.239178] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b91bbc5c-0654-40c2-bb5d-5df819e801d4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.248307] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c02c14cc-903b-4279-9294-6fa98f845aa2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1023.248603] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1f1ceda5-54ba-4bb6-bb90-61c1e402c51f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.256311] env[62385]: DEBUG oslo_vmware.api [None req-c02c14cc-903b-4279-9294-6fa98f845aa2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1023.256311] env[62385]: value = "task-1206358" [ 1023.256311] env[62385]: _type = "Task" [ 1023.256311] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.265119] env[62385]: DEBUG oslo_vmware.api [None req-c02c14cc-903b-4279-9294-6fa98f845aa2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206358, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.303750] env[62385]: DEBUG oslo_vmware.api [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206357, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.363471] env[62385]: INFO nova.compute.manager [None req-db9034be-4bf6-4cb8-a612-9f11df348d57 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Detaching volume 9621f12c-980a-4762-8ea2-c4fff11dc20e [ 1023.376014] env[62385]: DEBUG nova.network.neutron [-] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.413932] env[62385]: INFO nova.virt.block_device [None req-db9034be-4bf6-4cb8-a612-9f11df348d57 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Attempting to driver detach volume 9621f12c-980a-4762-8ea2-c4fff11dc20e from mountpoint /dev/sdb [ 1023.414343] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-db9034be-4bf6-4cb8-a612-9f11df348d57 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Volume detach. Driver type: vmdk {{(pid=62385) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1023.414635] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-db9034be-4bf6-4cb8-a612-9f11df348d57 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261287', 'volume_id': '9621f12c-980a-4762-8ea2-c4fff11dc20e', 'name': 'volume-9621f12c-980a-4762-8ea2-c4fff11dc20e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5daf0a99-0c2a-40d8-afc7-1998e21e32d8', 'attached_at': '', 'detached_at': '', 'volume_id': '9621f12c-980a-4762-8ea2-c4fff11dc20e', 'serial': '9621f12c-980a-4762-8ea2-c4fff11dc20e'} {{(pid=62385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1023.415949] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8b3c952-f15b-485f-8005-f2b2e2f28807 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.444037] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c28c3305-7259-493e-b8e7-3c4fb25114bc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.452358] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a957c1bc-f980-404b-92e2-e372c1932002 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.473843] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef916234-8217-48c4-b7e6-b9bc374983c7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.490729] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-db9034be-4bf6-4cb8-a612-9f11df348d57 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] The volume has not been displaced from its original location: [datastore1] volume-9621f12c-980a-4762-8ea2-c4fff11dc20e/volume-9621f12c-980a-4762-8ea2-c4fff11dc20e.vmdk. No consolidation needed. {{(pid=62385) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1023.495781] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-db9034be-4bf6-4cb8-a612-9f11df348d57 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Reconfiguring VM instance instance-0000005d to detach disk 2001 {{(pid=62385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1023.496205] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-70562eaa-5253-4549-9859-a5d9e0940bd3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.515518] env[62385]: DEBUG oslo_vmware.api [None req-db9034be-4bf6-4cb8-a612-9f11df348d57 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1023.515518] env[62385]: value = "task-1206359" [ 1023.515518] env[62385]: _type = "Task" [ 1023.515518] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.525575] env[62385]: DEBUG oslo_vmware.api [None req-db9034be-4bf6-4cb8-a612-9f11df348d57 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206359, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.599448] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.311s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.599599] env[62385]: DEBUG nova.compute.manager [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1023.602639] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 4.552s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.703366] env[62385]: DEBUG oslo_concurrency.lockutils [req-a9116891-734d-4168-b799-f8187fb5d0f2 req-fe728fd6-f7e8-48aa-a792-d25fe2412259 service nova] Releasing lock "refresh_cache-47311a9c-74d2-400d-a22d-9ef55b14edb5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1023.703785] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquired lock "refresh_cache-47311a9c-74d2-400d-a22d-9ef55b14edb5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.703949] env[62385]: DEBUG nova.network.neutron [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1023.766650] env[62385]: DEBUG oslo_vmware.api [None req-c02c14cc-903b-4279-9294-6fa98f845aa2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206358, 'name': PowerOffVM_Task, 'duration_secs': 0.406739} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.766917] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c02c14cc-903b-4279-9294-6fa98f845aa2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1023.767115] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c02c14cc-903b-4279-9294-6fa98f845aa2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1023.767378] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-993dd5e2-7647-4c0a-b4b6-265538aa5c50 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.804114] env[62385]: DEBUG oslo_vmware.api [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206357, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.821514] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-57468fc3-c541-446a-91b3-64d71c7db5a6 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Volume attach. Driver type: vmdk {{(pid=62385) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1023.821748] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-57468fc3-c541-446a-91b3-64d71c7db5a6 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261293', 'volume_id': '3b02b9da-5324-4812-abd7-09ff15288694', 'name': 'volume-3b02b9da-5324-4812-abd7-09ff15288694', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a1c71d8b-0953-4cc2-961d-0c2b2fc11110', 'attached_at': '', 'detached_at': '', 'volume_id': '3b02b9da-5324-4812-abd7-09ff15288694', 'serial': '3b02b9da-5324-4812-abd7-09ff15288694'} {{(pid=62385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1023.822695] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99a92752-2c6e-4e5e-b811-5560feebccef {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.838269] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-291deb1a-aed7-4548-b171-f7d15adfc18b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.865294] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-57468fc3-c541-446a-91b3-64d71c7db5a6 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] volume-3b02b9da-5324-4812-abd7-09ff15288694/volume-3b02b9da-5324-4812-abd7-09ff15288694.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1023.865595] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-82802b7b-d5e3-479a-9b33-1d6b42b6e44a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.878427] env[62385]: INFO nova.compute.manager [-] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Took 1.32 seconds to deallocate network for instance. [ 1023.885958] env[62385]: DEBUG oslo_vmware.api [None req-57468fc3-c541-446a-91b3-64d71c7db5a6 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for the task: (returnval){ [ 1023.885958] env[62385]: value = "task-1206361" [ 1023.885958] env[62385]: _type = "Task" [ 1023.885958] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.894491] env[62385]: DEBUG oslo_vmware.api [None req-57468fc3-c541-446a-91b3-64d71c7db5a6 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206361, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.027500] env[62385]: DEBUG oslo_vmware.api [None req-db9034be-4bf6-4cb8-a612-9f11df348d57 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206359, 'name': ReconfigVM_Task, 'duration_secs': 0.318136} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.027940] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-db9034be-4bf6-4cb8-a612-9f11df348d57 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Reconfigured VM instance instance-0000005d to detach disk 2001 {{(pid=62385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1024.034227] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a7ae2c08-a41f-48dd-8719-a4c758d8dd4f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.045638] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c02c14cc-903b-4279-9294-6fa98f845aa2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1024.045902] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c02c14cc-903b-4279-9294-6fa98f845aa2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1024.046155] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c02c14cc-903b-4279-9294-6fa98f845aa2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Deleting the datastore file [datastore1] 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1024.046905] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6a5efe66-7eec-43f6-a020-b447753b5d6f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.056313] env[62385]: DEBUG oslo_vmware.api [None req-db9034be-4bf6-4cb8-a612-9f11df348d57 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1024.056313] env[62385]: value = "task-1206363" [ 1024.056313] env[62385]: _type = "Task" [ 1024.056313] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.056760] env[62385]: DEBUG oslo_vmware.api [None req-c02c14cc-903b-4279-9294-6fa98f845aa2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1024.056760] env[62385]: value = "task-1206362" [ 1024.056760] env[62385]: _type = "Task" [ 1024.056760] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.080169] env[62385]: DEBUG oslo_vmware.api [None req-c02c14cc-903b-4279-9294-6fa98f845aa2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206362, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.080496] env[62385]: DEBUG oslo_vmware.api [None req-db9034be-4bf6-4cb8-a612-9f11df348d57 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206363, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.107274] env[62385]: DEBUG nova.compute.utils [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1024.117502] env[62385]: DEBUG nova.compute.manager [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1024.117741] env[62385]: DEBUG nova.network.neutron [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1024.227702] env[62385]: DEBUG nova.policy [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e234024cef314157add78767354b5929', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '58f5feadcb364e16bf9e23e8c6990318', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 1024.262160] env[62385]: DEBUG nova.network.neutron [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1024.304151] env[62385]: DEBUG oslo_vmware.api [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206357, 'name': ReconfigVM_Task, 'duration_secs': 1.028855} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.304444] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Reconfigured VM instance instance-00000068 to attach disk [datastore1] 68748d25-017f-4b58-9984-70264abbc5ba/68748d25-017f-4b58-9984-70264abbc5ba.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1024.305102] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-236c43cd-f941-4933-b27f-575e7167fb2b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.312588] env[62385]: DEBUG oslo_vmware.api [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 1024.312588] env[62385]: value = "task-1206364" [ 1024.312588] env[62385]: _type = "Task" [ 1024.312588] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.327825] env[62385]: DEBUG oslo_vmware.api [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206364, 'name': Rename_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.386129] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c6808a79-c5d5-48a7-9431-20da78d448b1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.395879] env[62385]: DEBUG oslo_vmware.api [None req-57468fc3-c541-446a-91b3-64d71c7db5a6 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206361, 'name': ReconfigVM_Task, 'duration_secs': 0.396062} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.396185] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-57468fc3-c541-446a-91b3-64d71c7db5a6 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Reconfigured VM instance instance-00000063 to attach disk [datastore1] volume-3b02b9da-5324-4812-abd7-09ff15288694/volume-3b02b9da-5324-4812-abd7-09ff15288694.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1024.400966] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-19704858-7db7-41d4-b919-a6846a9b0d1c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.416971] env[62385]: DEBUG oslo_vmware.api [None req-57468fc3-c541-446a-91b3-64d71c7db5a6 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for the task: (returnval){ [ 1024.416971] env[62385]: value = "task-1206365" [ 1024.416971] env[62385]: _type = "Task" [ 1024.416971] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.425578] env[62385]: DEBUG oslo_vmware.api [None req-57468fc3-c541-446a-91b3-64d71c7db5a6 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206365, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.569746] env[62385]: DEBUG oslo_vmware.api [None req-db9034be-4bf6-4cb8-a612-9f11df348d57 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206363, 'name': ReconfigVM_Task, 'duration_secs': 0.164525} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.570547] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-db9034be-4bf6-4cb8-a612-9f11df348d57 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261287', 'volume_id': '9621f12c-980a-4762-8ea2-c4fff11dc20e', 'name': 'volume-9621f12c-980a-4762-8ea2-c4fff11dc20e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5daf0a99-0c2a-40d8-afc7-1998e21e32d8', 'attached_at': '', 'detached_at': '', 'volume_id': '9621f12c-980a-4762-8ea2-c4fff11dc20e', 'serial': '9621f12c-980a-4762-8ea2-c4fff11dc20e'} {{(pid=62385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1024.576125] env[62385]: DEBUG oslo_vmware.api [None req-c02c14cc-903b-4279-9294-6fa98f845aa2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206362, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.202168} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.576417] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c02c14cc-903b-4279-9294-6fa98f845aa2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1024.576604] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c02c14cc-903b-4279-9294-6fa98f845aa2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1024.576797] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c02c14cc-903b-4279-9294-6fa98f845aa2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1024.576979] env[62385]: INFO nova.compute.manager [None req-c02c14cc-903b-4279-9294-6fa98f845aa2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Took 1.34 seconds to destroy the instance on the hypervisor. [ 1024.577239] env[62385]: DEBUG oslo.service.loopingcall [None req-c02c14cc-903b-4279-9294-6fa98f845aa2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1024.578531] env[62385]: DEBUG nova.compute.manager [-] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1024.578531] env[62385]: DEBUG nova.network.neutron [-] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1024.601456] env[62385]: DEBUG nova.network.neutron [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Updating instance_info_cache with network_info: [{"id": "4af1e8e6-edeb-47ff-b1b1-f4f974a69544", "address": "fa:16:3e:74:09:51", "network": {"id": "8b2d7a6e-25d4-4a54-ab54-5972612046da", "bridge": "br-int", "label": "tempest-ServersTestJSON-1142134211-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bf8e3c942d3445919cfbe988cca84e90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4af1e8e6-ed", "ovs_interfaceid": "4af1e8e6-edeb-47ff-b1b1-f4f974a69544", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.618846] env[62385]: DEBUG nova.compute.manager [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1024.651453] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 9d6f098a-0b05-43ef-96b0-9eb99ad3538c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1024.651796] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 5daf0a99-0c2a-40d8-afc7-1998e21e32d8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1024.652065] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance a1c71d8b-0953-4cc2-961d-0c2b2fc11110 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1024.652519] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 57637d6e-39e2-405c-bbcf-64bf0ffbebd1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1024.652655] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1024.652778] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance b0a6a3b9-05e4-4190-9cc6-4237dee07f91 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1024.652899] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 68748d25-017f-4b58-9984-70264abbc5ba actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1024.763377] env[62385]: DEBUG nova.compute.manager [req-6bd51825-2f7b-4d61-9d25-79ac2b3a11cb req-aa96f374-1a1c-422d-8270-5486f897bbe8 service nova] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Received event network-vif-deleted-c8bba88c-566a-4a08-9526-4365303ef423 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1024.825211] env[62385]: DEBUG oslo_vmware.api [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206364, 'name': Rename_Task, 'duration_secs': 0.327993} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.825531] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1024.825667] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-affccf0c-0e61-443a-99e0-484d468c8a8c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.833158] env[62385]: DEBUG oslo_vmware.api [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 1024.833158] env[62385]: value = "task-1206366" [ 1024.833158] env[62385]: _type = "Task" [ 1024.833158] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.841540] env[62385]: DEBUG oslo_vmware.api [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206366, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.927391] env[62385]: DEBUG oslo_vmware.api [None req-57468fc3-c541-446a-91b3-64d71c7db5a6 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206365, 'name': ReconfigVM_Task, 'duration_secs': 0.153254} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.927812] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-57468fc3-c541-446a-91b3-64d71c7db5a6 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261293', 'volume_id': '3b02b9da-5324-4812-abd7-09ff15288694', 'name': 'volume-3b02b9da-5324-4812-abd7-09ff15288694', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a1c71d8b-0953-4cc2-961d-0c2b2fc11110', 'attached_at': '', 'detached_at': '', 'volume_id': '3b02b9da-5324-4812-abd7-09ff15288694', 'serial': '3b02b9da-5324-4812-abd7-09ff15288694'} {{(pid=62385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1025.103733] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Releasing lock "refresh_cache-47311a9c-74d2-400d-a22d-9ef55b14edb5" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.104093] env[62385]: DEBUG nova.compute.manager [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Instance network_info: |[{"id": "4af1e8e6-edeb-47ff-b1b1-f4f974a69544", "address": "fa:16:3e:74:09:51", "network": {"id": "8b2d7a6e-25d4-4a54-ab54-5972612046da", "bridge": "br-int", "label": "tempest-ServersTestJSON-1142134211-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bf8e3c942d3445919cfbe988cca84e90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1", "external-id": "nsx-vlan-transportzone-880", "segmentation_id": 880, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4af1e8e6-ed", "ovs_interfaceid": "4af1e8e6-edeb-47ff-b1b1-f4f974a69544", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1025.104538] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:74:09:51', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8b1a0b17-d008-4a8b-be2a-796ff1a9a2d1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4af1e8e6-edeb-47ff-b1b1-f4f974a69544', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1025.113919] env[62385]: DEBUG oslo.service.loopingcall [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1025.114238] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1025.114485] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ccf15fb6-6695-401b-b01d-20607df7a993 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.132564] env[62385]: DEBUG nova.objects.instance [None req-db9034be-4bf6-4cb8-a612-9f11df348d57 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lazy-loading 'flavor' on Instance uuid 5daf0a99-0c2a-40d8-afc7-1998e21e32d8 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1025.135160] env[62385]: INFO nova.virt.block_device [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Booting with volume e44d7ff8-cd5f-4a36-a788-a75683085f24 at /dev/sda [ 1025.143845] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1025.143845] env[62385]: value = "task-1206367" [ 1025.143845] env[62385]: _type = "Task" [ 1025.143845] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.157332] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206367, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.160480] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 8e1d416c-2624-468c-94d7-6265e9f4178f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1025.160480] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 47311a9c-74d2-400d-a22d-9ef55b14edb5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1025.160480] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1025.160480] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Total usable vcpus: 48, total allocated vcpus: 9 {{(pid=62385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1025.160480] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2304MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=9 pci_stats=[] {{(pid=62385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1025.177022] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-08261e9b-a5de-46ab-9e56-398e4d518b83 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.189765] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad0997fd-e7ef-46db-b49c-8ee0edae4676 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.227715] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-87b92032-6bfd-41d5-b738-c128621578f8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.239029] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5411627e-4563-4867-ac4d-699d093c2714 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.278782] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-924a56ec-8416-41cd-9f04-6a358092c8f7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.287384] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8ac8be8-8240-4580-b36a-93fd93b5bf98 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.304754] env[62385]: DEBUG nova.virt.block_device [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Updating existing volume attachment record: d9397208-d7e1-48ca-a529-fd567e4c0ab0 {{(pid=62385) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1025.345612] env[62385]: DEBUG oslo_vmware.api [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206366, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.381085] env[62385]: DEBUG nova.network.neutron [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Successfully created port: 5af3050a-7535-42c3-bb8d-4f1141049e32 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1025.402825] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57faac15-1de6-4b4b-93a7-2e75fa6a1b7e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.412203] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74c4f7a4-ca97-4462-8086-a72330bc8f68 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.455410] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93f7cfd2-76d9-4ea1-b1e9-3c5e605bd813 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.465833] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0e2f940-4948-43bf-9a31-c2bafa8f9553 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.483175] env[62385]: DEBUG nova.compute.provider_tree [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1025.661573] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206367, 'name': CreateVM_Task, 'duration_secs': 0.344264} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.662397] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1025.663844] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.663961] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.664336] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1025.664940] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8655f8a9-6de2-4e15-9a8e-1aff3d5b548d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.675317] env[62385]: DEBUG oslo_vmware.api [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 1025.675317] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c260df-dd44-9e90-6878-2748072f550b" [ 1025.675317] env[62385]: _type = "Task" [ 1025.675317] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.685419] env[62385]: DEBUG oslo_vmware.api [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c260df-dd44-9e90-6878-2748072f550b, 'name': SearchDatastore_Task, 'duration_secs': 0.010227} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.685733] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.685975] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1025.686245] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.686393] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.686570] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1025.686831] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d16976f8-5444-47f8-b5a5-f039328b24f9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.696137] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1025.696205] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1025.696934] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7293436c-2847-41d2-acc4-9765e0ce1f5b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.705203] env[62385]: DEBUG oslo_vmware.api [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 1025.705203] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]526974c5-c9fc-3a97-7e20-cef66076d170" [ 1025.705203] env[62385]: _type = "Task" [ 1025.705203] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.714121] env[62385]: DEBUG oslo_vmware.api [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]526974c5-c9fc-3a97-7e20-cef66076d170, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.844819] env[62385]: DEBUG oslo_vmware.api [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206366, 'name': PowerOnVM_Task, 'duration_secs': 0.708824} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.845203] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1025.845924] env[62385]: INFO nova.compute.manager [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Took 8.93 seconds to spawn the instance on the hypervisor. [ 1025.845924] env[62385]: DEBUG nova.compute.manager [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1025.846465] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1630bad4-e726-48f7-8050-5d9bb0505d04 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.988939] env[62385]: DEBUG nova.scheduler.client.report [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1026.006476] env[62385]: DEBUG nova.objects.instance [None req-57468fc3-c541-446a-91b3-64d71c7db5a6 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lazy-loading 'flavor' on Instance uuid a1c71d8b-0953-4cc2-961d-0c2b2fc11110 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1026.142805] env[62385]: DEBUG oslo_concurrency.lockutils [None req-db9034be-4bf6-4cb8-a612-9f11df348d57 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "5daf0a99-0c2a-40d8-afc7-1998e21e32d8" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.282s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.215066] env[62385]: DEBUG oslo_vmware.api [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]526974c5-c9fc-3a97-7e20-cef66076d170, 'name': SearchDatastore_Task, 'duration_secs': 0.009784} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.216037] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-673f0e92-f5f6-4626-8b9f-94146fd8d68a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.223524] env[62385]: DEBUG oslo_vmware.api [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 1026.223524] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]522540bc-f29e-533e-b546-7bf4544b6278" [ 1026.223524] env[62385]: _type = "Task" [ 1026.223524] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.234812] env[62385]: DEBUG oslo_vmware.api [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]522540bc-f29e-533e-b546-7bf4544b6278, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.258027] env[62385]: DEBUG nova.network.neutron [-] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.367524] env[62385]: INFO nova.compute.manager [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Took 18.21 seconds to build instance. [ 1026.394808] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9a98fbb0-158e-4d6e-95e6-91542f6881e4 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "5daf0a99-0c2a-40d8-afc7-1998e21e32d8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.394808] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9a98fbb0-158e-4d6e-95e6-91542f6881e4 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "5daf0a99-0c2a-40d8-afc7-1998e21e32d8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.394808] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9a98fbb0-158e-4d6e-95e6-91542f6881e4 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "5daf0a99-0c2a-40d8-afc7-1998e21e32d8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.395298] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9a98fbb0-158e-4d6e-95e6-91542f6881e4 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "5daf0a99-0c2a-40d8-afc7-1998e21e32d8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.395298] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9a98fbb0-158e-4d6e-95e6-91542f6881e4 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "5daf0a99-0c2a-40d8-afc7-1998e21e32d8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.397698] env[62385]: INFO nova.compute.manager [None req-9a98fbb0-158e-4d6e-95e6-91542f6881e4 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Terminating instance [ 1026.399377] env[62385]: DEBUG nova.compute.manager [None req-9a98fbb0-158e-4d6e-95e6-91542f6881e4 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1026.399623] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-9a98fbb0-158e-4d6e-95e6-91542f6881e4 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1026.400558] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddd06c86-7d87-465b-9595-172dba26b347 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.409455] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a98fbb0-158e-4d6e-95e6-91542f6881e4 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1026.409749] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-56d98018-a961-418e-926e-1eb2bac936e6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.416707] env[62385]: DEBUG oslo_vmware.api [None req-9a98fbb0-158e-4d6e-95e6-91542f6881e4 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1026.416707] env[62385]: value = "task-1206368" [ 1026.416707] env[62385]: _type = "Task" [ 1026.416707] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.425268] env[62385]: DEBUG oslo_vmware.api [None req-9a98fbb0-158e-4d6e-95e6-91542f6881e4 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206368, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.498842] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62385) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1026.499089] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.897s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.499411] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.107s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.499623] env[62385]: DEBUG nova.objects.instance [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lazy-loading 'pci_requests' on Instance uuid 8e1d416c-2624-468c-94d7-6265e9f4178f {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1026.511598] env[62385]: DEBUG oslo_concurrency.lockutils [None req-57468fc3-c541-446a-91b3-64d71c7db5a6 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "a1c71d8b-0953-4cc2-961d-0c2b2fc11110" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.299s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.738697] env[62385]: DEBUG oslo_vmware.api [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]522540bc-f29e-533e-b546-7bf4544b6278, 'name': SearchDatastore_Task, 'duration_secs': 0.010659} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.738988] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.739278] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 47311a9c-74d2-400d-a22d-9ef55b14edb5/47311a9c-74d2-400d-a22d-9ef55b14edb5.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1026.739559] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bfaec6b7-410b-46c9-a027-b0f24692ba43 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.747142] env[62385]: DEBUG oslo_vmware.api [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 1026.747142] env[62385]: value = "task-1206369" [ 1026.747142] env[62385]: _type = "Task" [ 1026.747142] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.756414] env[62385]: DEBUG oslo_vmware.api [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206369, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.761455] env[62385]: INFO nova.compute.manager [-] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Took 2.18 seconds to deallocate network for instance. [ 1026.793094] env[62385]: DEBUG nova.compute.manager [req-931a98f0-b7a9-49c4-b07c-8f23ea57adcd req-77a8d97d-30a0-4037-9d3d-9b3ad0be061a service nova] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Received event network-vif-deleted-404c104c-258f-480e-bfe2-4ace4e83a43c {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1026.871052] env[62385]: DEBUG oslo_concurrency.lockutils [None req-eb5e7fac-ab2a-49e5-aa19-91e36e3cc098 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "68748d25-017f-4b58-9984-70264abbc5ba" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.725s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.936336] env[62385]: DEBUG oslo_vmware.api [None req-9a98fbb0-158e-4d6e-95e6-91542f6881e4 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206368, 'name': PowerOffVM_Task, 'duration_secs': 0.283631} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.942019] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a98fbb0-158e-4d6e-95e6-91542f6881e4 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1026.942019] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-9a98fbb0-158e-4d6e-95e6-91542f6881e4 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1026.942019] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-272791bd-d68a-4872-ab7f-6a9b76ad34d6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.004351] env[62385]: DEBUG nova.objects.instance [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lazy-loading 'numa_topology' on Instance uuid 8e1d416c-2624-468c-94d7-6265e9f4178f {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1027.024917] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-9a98fbb0-158e-4d6e-95e6-91542f6881e4 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1027.024917] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-9a98fbb0-158e-4d6e-95e6-91542f6881e4 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1027.024917] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a98fbb0-158e-4d6e-95e6-91542f6881e4 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Deleting the datastore file [datastore1] 5daf0a99-0c2a-40d8-afc7-1998e21e32d8 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1027.024917] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bb91ee6b-657d-4c5b-bce7-cea754f2dc71 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.032546] env[62385]: DEBUG oslo_vmware.api [None req-9a98fbb0-158e-4d6e-95e6-91542f6881e4 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1027.032546] env[62385]: value = "task-1206371" [ 1027.032546] env[62385]: _type = "Task" [ 1027.032546] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.043030] env[62385]: DEBUG oslo_vmware.api [None req-9a98fbb0-158e-4d6e-95e6-91542f6881e4 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206371, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.257659] env[62385]: DEBUG oslo_vmware.api [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206369, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.473015} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.258394] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 47311a9c-74d2-400d-a22d-9ef55b14edb5/47311a9c-74d2-400d-a22d-9ef55b14edb5.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1027.258533] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1027.258774] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d78105d0-f321-4dca-ae97-21e5f0cb7857 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.267492] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c02c14cc-903b-4279-9294-6fa98f845aa2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.267969] env[62385]: DEBUG oslo_vmware.api [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 1027.267969] env[62385]: value = "task-1206372" [ 1027.267969] env[62385]: _type = "Task" [ 1027.267969] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.277956] env[62385]: DEBUG oslo_vmware.api [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206372, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.377723] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Acquiring lock "c49d4366-2ea0-44f2-8463-daf62bd1c40c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.377723] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Lock "c49d4366-2ea0-44f2-8463-daf62bd1c40c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.424307] env[62385]: DEBUG oslo_concurrency.lockutils [None req-782c0473-e437-423d-bdca-2942fb468e6d tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Acquiring lock "a1c71d8b-0953-4cc2-961d-0c2b2fc11110" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.424574] env[62385]: DEBUG oslo_concurrency.lockutils [None req-782c0473-e437-423d-bdca-2942fb468e6d tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "a1c71d8b-0953-4cc2-961d-0c2b2fc11110" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.442252] env[62385]: DEBUG nova.compute.manager [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1027.442973] env[62385]: DEBUG nova.virt.hardware [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1027.443224] env[62385]: DEBUG nova.virt.hardware [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1027.443384] env[62385]: DEBUG nova.virt.hardware [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1027.443623] env[62385]: DEBUG nova.virt.hardware [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1027.443722] env[62385]: DEBUG nova.virt.hardware [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1027.443872] env[62385]: DEBUG nova.virt.hardware [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1027.444133] env[62385]: DEBUG nova.virt.hardware [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1027.444257] env[62385]: DEBUG nova.virt.hardware [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1027.444569] env[62385]: DEBUG nova.virt.hardware [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1027.444631] env[62385]: DEBUG nova.virt.hardware [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1027.444768] env[62385]: DEBUG nova.virt.hardware [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1027.446129] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a1b8601-558f-469a-b879-cd875adc3e59 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.449803] env[62385]: DEBUG nova.network.neutron [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Successfully updated port: 5af3050a-7535-42c3-bb8d-4f1141049e32 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1027.457646] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45076240-6cc4-4cbc-af31-85674da03c0f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.507441] env[62385]: INFO nova.compute.claims [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1027.548094] env[62385]: DEBUG oslo_vmware.api [None req-9a98fbb0-158e-4d6e-95e6-91542f6881e4 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206371, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.289716} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.548420] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a98fbb0-158e-4d6e-95e6-91542f6881e4 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1027.549421] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-9a98fbb0-158e-4d6e-95e6-91542f6881e4 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1027.549421] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-9a98fbb0-158e-4d6e-95e6-91542f6881e4 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1027.549421] env[62385]: INFO nova.compute.manager [None req-9a98fbb0-158e-4d6e-95e6-91542f6881e4 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1027.549421] env[62385]: DEBUG oslo.service.loopingcall [None req-9a98fbb0-158e-4d6e-95e6-91542f6881e4 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1027.549609] env[62385]: DEBUG nova.compute.manager [-] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1027.549609] env[62385]: DEBUG nova.network.neutron [-] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1027.663244] env[62385]: DEBUG nova.compute.manager [req-1b2075be-0562-4ac5-86f3-b31b2065b873 req-09683aaa-b85e-4ac0-882b-1a7d1efc9a6a service nova] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Received event network-changed-a7153a2d-db69-4d14-8ddc-b634ddabbc18 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1027.664220] env[62385]: DEBUG nova.compute.manager [req-1b2075be-0562-4ac5-86f3-b31b2065b873 req-09683aaa-b85e-4ac0-882b-1a7d1efc9a6a service nova] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Refreshing instance network info cache due to event network-changed-a7153a2d-db69-4d14-8ddc-b634ddabbc18. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1027.664662] env[62385]: DEBUG oslo_concurrency.lockutils [req-1b2075be-0562-4ac5-86f3-b31b2065b873 req-09683aaa-b85e-4ac0-882b-1a7d1efc9a6a service nova] Acquiring lock "refresh_cache-57637d6e-39e2-405c-bbcf-64bf0ffbebd1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.666711] env[62385]: DEBUG oslo_concurrency.lockutils [req-1b2075be-0562-4ac5-86f3-b31b2065b873 req-09683aaa-b85e-4ac0-882b-1a7d1efc9a6a service nova] Acquired lock "refresh_cache-57637d6e-39e2-405c-bbcf-64bf0ffbebd1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.666711] env[62385]: DEBUG nova.network.neutron [req-1b2075be-0562-4ac5-86f3-b31b2065b873 req-09683aaa-b85e-4ac0-882b-1a7d1efc9a6a service nova] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Refreshing network info cache for port a7153a2d-db69-4d14-8ddc-b634ddabbc18 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1027.780737] env[62385]: DEBUG oslo_vmware.api [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206372, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082683} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.781267] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1027.782514] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-185e03e3-c49c-439f-8c4e-dc4c79a068dd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.812997] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] 47311a9c-74d2-400d-a22d-9ef55b14edb5/47311a9c-74d2-400d-a22d-9ef55b14edb5.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1027.814105] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-beb1c817-8d79-4dec-bba0-411dc8b1d7d4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.841025] env[62385]: DEBUG oslo_vmware.api [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 1027.841025] env[62385]: value = "task-1206373" [ 1027.841025] env[62385]: _type = "Task" [ 1027.841025] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.849185] env[62385]: DEBUG oslo_vmware.api [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206373, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.880159] env[62385]: DEBUG nova.compute.manager [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1027.929677] env[62385]: DEBUG nova.compute.utils [None req-782c0473-e437-423d-bdca-2942fb468e6d tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1027.952570] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Acquiring lock "refresh_cache-5a5bcb7c-86cd-4950-a3dd-15515ae0aca6" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.954394] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Acquired lock "refresh_cache-5a5bcb7c-86cd-4950-a3dd-15515ae0aca6" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.954394] env[62385]: DEBUG nova.network.neutron [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1028.351100] env[62385]: DEBUG oslo_vmware.api [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206373, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.416658] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.437789] env[62385]: DEBUG oslo_concurrency.lockutils [None req-782c0473-e437-423d-bdca-2942fb468e6d tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "a1c71d8b-0953-4cc2-961d-0c2b2fc11110" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.013s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.518430] env[62385]: DEBUG nova.network.neutron [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1028.557641] env[62385]: DEBUG nova.network.neutron [req-1b2075be-0562-4ac5-86f3-b31b2065b873 req-09683aaa-b85e-4ac0-882b-1a7d1efc9a6a service nova] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Updated VIF entry in instance network info cache for port a7153a2d-db69-4d14-8ddc-b634ddabbc18. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1028.560360] env[62385]: DEBUG nova.network.neutron [req-1b2075be-0562-4ac5-86f3-b31b2065b873 req-09683aaa-b85e-4ac0-882b-1a7d1efc9a6a service nova] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Updating instance_info_cache with network_info: [{"id": "a7153a2d-db69-4d14-8ddc-b634ddabbc18", "address": "fa:16:3e:a3:0d:ab", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7153a2d-db", "ovs_interfaceid": "a7153a2d-db69-4d14-8ddc-b634ddabbc18", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.725319] env[62385]: DEBUG nova.network.neutron [-] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.733385] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54affe92-1a27-4ccb-a91c-cbc72c3c62ef {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.743118] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7ec5592-1cae-49f0-9d8c-263329cfed47 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.775450] env[62385]: DEBUG nova.network.neutron [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Updating instance_info_cache with network_info: [{"id": "5af3050a-7535-42c3-bb8d-4f1141049e32", "address": "fa:16:3e:3d:8b:5f", "network": {"id": "b428b89a-516a-4b8b-9a77-8d922c1865aa", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-105102189-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58f5feadcb364e16bf9e23e8c6990318", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "62f28d75-4e6a-4ae5-b8b3-d0652ea26d08", "external-id": "nsx-vlan-transportzone-801", "segmentation_id": 801, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5af3050a-75", "ovs_interfaceid": "5af3050a-7535-42c3-bb8d-4f1141049e32", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.778228] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d9577cd-029f-44a7-879b-8ffe670f265b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.791097] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9473e2cf-9a93-42dd-b3d0-ef4f0d789323 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.807239] env[62385]: DEBUG nova.compute.provider_tree [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1028.849568] env[62385]: DEBUG oslo_vmware.api [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206373, 'name': ReconfigVM_Task, 'duration_secs': 0.664546} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.850289] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Reconfigured VM instance instance-00000069 to attach disk [datastore1] 47311a9c-74d2-400d-a22d-9ef55b14edb5/47311a9c-74d2-400d-a22d-9ef55b14edb5.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1028.850902] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-11488458-5912-454e-a1ff-96911c8f786f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.859221] env[62385]: DEBUG oslo_vmware.api [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 1028.859221] env[62385]: value = "task-1206374" [ 1028.859221] env[62385]: _type = "Task" [ 1028.859221] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.872154] env[62385]: DEBUG oslo_vmware.api [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206374, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.928481] env[62385]: DEBUG nova.compute.manager [req-08ec0fab-0cb2-4c35-b7bb-eaf25e835a1a req-063d53e0-fc0f-4aa3-96aa-97c03b6a4e0a service nova] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Received event network-vif-plugged-5af3050a-7535-42c3-bb8d-4f1141049e32 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1028.928481] env[62385]: DEBUG oslo_concurrency.lockutils [req-08ec0fab-0cb2-4c35-b7bb-eaf25e835a1a req-063d53e0-fc0f-4aa3-96aa-97c03b6a4e0a service nova] Acquiring lock "5a5bcb7c-86cd-4950-a3dd-15515ae0aca6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.928481] env[62385]: DEBUG oslo_concurrency.lockutils [req-08ec0fab-0cb2-4c35-b7bb-eaf25e835a1a req-063d53e0-fc0f-4aa3-96aa-97c03b6a4e0a service nova] Lock "5a5bcb7c-86cd-4950-a3dd-15515ae0aca6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.928862] env[62385]: DEBUG oslo_concurrency.lockutils [req-08ec0fab-0cb2-4c35-b7bb-eaf25e835a1a req-063d53e0-fc0f-4aa3-96aa-97c03b6a4e0a service nova] Lock "5a5bcb7c-86cd-4950-a3dd-15515ae0aca6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.928862] env[62385]: DEBUG nova.compute.manager [req-08ec0fab-0cb2-4c35-b7bb-eaf25e835a1a req-063d53e0-fc0f-4aa3-96aa-97c03b6a4e0a service nova] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] No waiting events found dispatching network-vif-plugged-5af3050a-7535-42c3-bb8d-4f1141049e32 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1028.928950] env[62385]: WARNING nova.compute.manager [req-08ec0fab-0cb2-4c35-b7bb-eaf25e835a1a req-063d53e0-fc0f-4aa3-96aa-97c03b6a4e0a service nova] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Received unexpected event network-vif-plugged-5af3050a-7535-42c3-bb8d-4f1141049e32 for instance with vm_state building and task_state spawning. [ 1028.929127] env[62385]: DEBUG nova.compute.manager [req-08ec0fab-0cb2-4c35-b7bb-eaf25e835a1a req-063d53e0-fc0f-4aa3-96aa-97c03b6a4e0a service nova] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Received event network-changed-5af3050a-7535-42c3-bb8d-4f1141049e32 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1028.929248] env[62385]: DEBUG nova.compute.manager [req-08ec0fab-0cb2-4c35-b7bb-eaf25e835a1a req-063d53e0-fc0f-4aa3-96aa-97c03b6a4e0a service nova] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Refreshing instance network info cache due to event network-changed-5af3050a-7535-42c3-bb8d-4f1141049e32. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1028.929633] env[62385]: DEBUG oslo_concurrency.lockutils [req-08ec0fab-0cb2-4c35-b7bb-eaf25e835a1a req-063d53e0-fc0f-4aa3-96aa-97c03b6a4e0a service nova] Acquiring lock "refresh_cache-5a5bcb7c-86cd-4950-a3dd-15515ae0aca6" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.062486] env[62385]: DEBUG oslo_concurrency.lockutils [req-1b2075be-0562-4ac5-86f3-b31b2065b873 req-09683aaa-b85e-4ac0-882b-1a7d1efc9a6a service nova] Releasing lock "refresh_cache-57637d6e-39e2-405c-bbcf-64bf0ffbebd1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.227881] env[62385]: INFO nova.compute.manager [-] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Took 1.68 seconds to deallocate network for instance. [ 1029.285163] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Releasing lock "refresh_cache-5a5bcb7c-86cd-4950-a3dd-15515ae0aca6" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.285499] env[62385]: DEBUG nova.compute.manager [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Instance network_info: |[{"id": "5af3050a-7535-42c3-bb8d-4f1141049e32", "address": "fa:16:3e:3d:8b:5f", "network": {"id": "b428b89a-516a-4b8b-9a77-8d922c1865aa", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-105102189-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58f5feadcb364e16bf9e23e8c6990318", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "62f28d75-4e6a-4ae5-b8b3-d0652ea26d08", "external-id": "nsx-vlan-transportzone-801", "segmentation_id": 801, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5af3050a-75", "ovs_interfaceid": "5af3050a-7535-42c3-bb8d-4f1141049e32", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1029.285847] env[62385]: DEBUG oslo_concurrency.lockutils [req-08ec0fab-0cb2-4c35-b7bb-eaf25e835a1a req-063d53e0-fc0f-4aa3-96aa-97c03b6a4e0a service nova] Acquired lock "refresh_cache-5a5bcb7c-86cd-4950-a3dd-15515ae0aca6" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.286318] env[62385]: DEBUG nova.network.neutron [req-08ec0fab-0cb2-4c35-b7bb-eaf25e835a1a req-063d53e0-fc0f-4aa3-96aa-97c03b6a4e0a service nova] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Refreshing network info cache for port 5af3050a-7535-42c3-bb8d-4f1141049e32 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1029.288138] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3d:8b:5f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '62f28d75-4e6a-4ae5-b8b3-d0652ea26d08', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5af3050a-7535-42c3-bb8d-4f1141049e32', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1029.296029] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Creating folder: Project (58f5feadcb364e16bf9e23e8c6990318). Parent ref: group-v261107. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1029.298874] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-98c53547-16c9-45a4-a6e2-80ffa6d41f8e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.309711] env[62385]: DEBUG nova.scheduler.client.report [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1029.314210] env[62385]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 1029.314376] env[62385]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=62385) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 1029.315160] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Folder already exists: Project (58f5feadcb364e16bf9e23e8c6990318). Parent ref: group-v261107. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1029.315160] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Creating folder: Instances. Parent ref: group-v261278. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1029.315320] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5d6012e2-5eb0-4813-ac33-6b4735a4d39f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.325219] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Created folder: Instances in parent group-v261278. [ 1029.325481] env[62385]: DEBUG oslo.service.loopingcall [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1029.326288] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1029.326996] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8fa5a760-6c6c-430e-a52b-64579cb81a5d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.352311] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1029.352311] env[62385]: value = "task-1206377" [ 1029.352311] env[62385]: _type = "Task" [ 1029.352311] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.357136] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206377, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.370178] env[62385]: DEBUG oslo_vmware.api [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206374, 'name': Rename_Task, 'duration_secs': 0.16894} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.370178] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1029.370178] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-039e6831-a4b1-4464-9260-f58570830dca {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.378407] env[62385]: DEBUG oslo_vmware.api [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 1029.378407] env[62385]: value = "task-1206378" [ 1029.378407] env[62385]: _type = "Task" [ 1029.378407] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.387412] env[62385]: DEBUG oslo_vmware.api [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206378, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.518567] env[62385]: DEBUG oslo_concurrency.lockutils [None req-782c0473-e437-423d-bdca-2942fb468e6d tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Acquiring lock "a1c71d8b-0953-4cc2-961d-0c2b2fc11110" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.518911] env[62385]: DEBUG oslo_concurrency.lockutils [None req-782c0473-e437-423d-bdca-2942fb468e6d tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "a1c71d8b-0953-4cc2-961d-0c2b2fc11110" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.519102] env[62385]: INFO nova.compute.manager [None req-782c0473-e437-423d-bdca-2942fb468e6d tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Attaching volume 67e47839-a66b-4032-b1c4-93171c686c2f to /dev/sdc [ 1029.557870] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7537bd87-bf90-4997-9041-39b7739ef08a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.565870] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3d93e12-fc5d-4087-a912-d4bab59cbedd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.580892] env[62385]: DEBUG nova.virt.block_device [None req-782c0473-e437-423d-bdca-2942fb468e6d tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Updating existing volume attachment record: 20f9e0e7-f089-469b-adbf-09730c0d2339 {{(pid=62385) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1029.705144] env[62385]: DEBUG nova.compute.manager [req-e16c58c1-b687-48c7-9338-0820696c5f69 req-2dd4dd66-076b-488e-a02d-2aa3aafb17ff service nova] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Received event network-changed-c308701b-9b9e-4ff3-9a01-c27aabf54f0d {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1029.705144] env[62385]: DEBUG nova.compute.manager [req-e16c58c1-b687-48c7-9338-0820696c5f69 req-2dd4dd66-076b-488e-a02d-2aa3aafb17ff service nova] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Refreshing instance network info cache due to event network-changed-c308701b-9b9e-4ff3-9a01-c27aabf54f0d. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1029.705144] env[62385]: DEBUG oslo_concurrency.lockutils [req-e16c58c1-b687-48c7-9338-0820696c5f69 req-2dd4dd66-076b-488e-a02d-2aa3aafb17ff service nova] Acquiring lock "refresh_cache-68748d25-017f-4b58-9984-70264abbc5ba" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.705322] env[62385]: DEBUG oslo_concurrency.lockutils [req-e16c58c1-b687-48c7-9338-0820696c5f69 req-2dd4dd66-076b-488e-a02d-2aa3aafb17ff service nova] Acquired lock "refresh_cache-68748d25-017f-4b58-9984-70264abbc5ba" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.705489] env[62385]: DEBUG nova.network.neutron [req-e16c58c1-b687-48c7-9338-0820696c5f69 req-2dd4dd66-076b-488e-a02d-2aa3aafb17ff service nova] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Refreshing network info cache for port c308701b-9b9e-4ff3-9a01-c27aabf54f0d {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1029.740459] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9a98fbb0-158e-4d6e-95e6-91542f6881e4 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.818956] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.319s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.821237] env[62385]: DEBUG oslo_concurrency.lockutils [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 8.689s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.821451] env[62385]: DEBUG nova.objects.instance [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62385) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1029.859019] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206377, 'name': CreateVM_Task} progress is 25%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.860454] env[62385]: DEBUG nova.network.neutron [req-08ec0fab-0cb2-4c35-b7bb-eaf25e835a1a req-063d53e0-fc0f-4aa3-96aa-97c03b6a4e0a service nova] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Updated VIF entry in instance network info cache for port 5af3050a-7535-42c3-bb8d-4f1141049e32. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1029.860793] env[62385]: DEBUG nova.network.neutron [req-08ec0fab-0cb2-4c35-b7bb-eaf25e835a1a req-063d53e0-fc0f-4aa3-96aa-97c03b6a4e0a service nova] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Updating instance_info_cache with network_info: [{"id": "5af3050a-7535-42c3-bb8d-4f1141049e32", "address": "fa:16:3e:3d:8b:5f", "network": {"id": "b428b89a-516a-4b8b-9a77-8d922c1865aa", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-105102189-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58f5feadcb364e16bf9e23e8c6990318", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "62f28d75-4e6a-4ae5-b8b3-d0652ea26d08", "external-id": "nsx-vlan-transportzone-801", "segmentation_id": 801, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5af3050a-75", "ovs_interfaceid": "5af3050a-7535-42c3-bb8d-4f1141049e32", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.879098] env[62385]: INFO nova.network.neutron [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Updating port deafdeb9-f59c-48f4-873b-6e4b27d4a688 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1029.890606] env[62385]: DEBUG oslo_vmware.api [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206378, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.359555] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206377, 'name': CreateVM_Task, 'duration_secs': 0.945067} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.359959] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1030.360547] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'disk_bus': None, 'attachment_id': 'd9397208-d7e1-48ca-a529-fd567e4c0ab0', 'boot_index': 0, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261288', 'volume_id': 'e44d7ff8-cd5f-4a36-a788-a75683085f24', 'name': 'volume-e44d7ff8-cd5f-4a36-a788-a75683085f24', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5a5bcb7c-86cd-4950-a3dd-15515ae0aca6', 'attached_at': '', 'detached_at': '', 'volume_id': 'e44d7ff8-cd5f-4a36-a788-a75683085f24', 'serial': 'e44d7ff8-cd5f-4a36-a788-a75683085f24'}, 'delete_on_termination': True, 'mount_device': '/dev/sda', 'device_type': None, 'guest_format': None, 'volume_type': None}], 'swap': None} {{(pid=62385) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1030.360735] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Root volume attach. Driver type: vmdk {{(pid=62385) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1030.361645] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec360388-8658-4918-990e-145516295085 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.364376] env[62385]: DEBUG oslo_concurrency.lockutils [req-08ec0fab-0cb2-4c35-b7bb-eaf25e835a1a req-063d53e0-fc0f-4aa3-96aa-97c03b6a4e0a service nova] Releasing lock "refresh_cache-5a5bcb7c-86cd-4950-a3dd-15515ae0aca6" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.364963] env[62385]: DEBUG nova.compute.manager [req-08ec0fab-0cb2-4c35-b7bb-eaf25e835a1a req-063d53e0-fc0f-4aa3-96aa-97c03b6a4e0a service nova] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Received event network-vif-deleted-1b04305d-1213-4f55-9b69-6f15de71f576 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1030.370570] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80505325-5df3-41fe-86f9-2deccca4e2e5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.382323] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5696b501-6da7-402a-bba5-1a4df19c378f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.396079] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-1c3f070d-c708-492e-9e18-4db3e42928b4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.400261] env[62385]: DEBUG oslo_vmware.api [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206378, 'name': PowerOnVM_Task} progress is 90%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.404710] env[62385]: DEBUG oslo_vmware.api [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Waiting for the task: (returnval){ [ 1030.404710] env[62385]: value = "task-1206380" [ 1030.404710] env[62385]: _type = "Task" [ 1030.404710] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.414735] env[62385]: DEBUG oslo_vmware.api [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Task: {'id': task-1206380, 'name': RelocateVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.459309] env[62385]: DEBUG nova.network.neutron [req-e16c58c1-b687-48c7-9338-0820696c5f69 req-2dd4dd66-076b-488e-a02d-2aa3aafb17ff service nova] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Updated VIF entry in instance network info cache for port c308701b-9b9e-4ff3-9a01-c27aabf54f0d. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1030.462020] env[62385]: DEBUG nova.network.neutron [req-e16c58c1-b687-48c7-9338-0820696c5f69 req-2dd4dd66-076b-488e-a02d-2aa3aafb17ff service nova] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Updating instance_info_cache with network_info: [{"id": "c308701b-9b9e-4ff3-9a01-c27aabf54f0d", "address": "fa:16:3e:9a:ed:8c", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc308701b-9b", "ovs_interfaceid": "c308701b-9b9e-4ff3-9a01-c27aabf54f0d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.831139] env[62385]: DEBUG oslo_concurrency.lockutils [None req-65d72856-ff4a-4502-ac4b-a5e573dd8282 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.009s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.831776] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c6808a79-c5d5-48a7-9431-20da78d448b1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.446s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.832023] env[62385]: DEBUG nova.objects.instance [None req-c6808a79-c5d5-48a7-9431-20da78d448b1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lazy-loading 'resources' on Instance uuid b0a6a3b9-05e4-4190-9cc6-4237dee07f91 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1030.895012] env[62385]: DEBUG oslo_vmware.api [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206378, 'name': PowerOnVM_Task, 'duration_secs': 1.142392} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.895408] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1030.895626] env[62385]: INFO nova.compute.manager [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Took 9.91 seconds to spawn the instance on the hypervisor. [ 1030.895811] env[62385]: DEBUG nova.compute.manager [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1030.896575] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1390651e-3081-461d-a86a-58350a1a15e8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.913030] env[62385]: DEBUG oslo_vmware.api [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Task: {'id': task-1206380, 'name': RelocateVM_Task, 'duration_secs': 0.481831} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.913310] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Volume attach. Driver type: vmdk {{(pid=62385) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1030.913517] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261288', 'volume_id': 'e44d7ff8-cd5f-4a36-a788-a75683085f24', 'name': 'volume-e44d7ff8-cd5f-4a36-a788-a75683085f24', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5a5bcb7c-86cd-4950-a3dd-15515ae0aca6', 'attached_at': '', 'detached_at': '', 'volume_id': 'e44d7ff8-cd5f-4a36-a788-a75683085f24', 'serial': 'e44d7ff8-cd5f-4a36-a788-a75683085f24'} {{(pid=62385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1030.914885] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d06de4c8-daeb-4b29-a206-963b0ecb3a28 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.930449] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72744296-f6b8-40e8-880d-fb35e8247e94 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.953042] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] volume-e44d7ff8-cd5f-4a36-a788-a75683085f24/volume-e44d7ff8-cd5f-4a36-a788-a75683085f24.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1030.953730] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-99a0e278-df5b-45c1-a7d8-dad4af128b66 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.968285] env[62385]: DEBUG oslo_concurrency.lockutils [req-e16c58c1-b687-48c7-9338-0820696c5f69 req-2dd4dd66-076b-488e-a02d-2aa3aafb17ff service nova] Releasing lock "refresh_cache-68748d25-017f-4b58-9984-70264abbc5ba" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.968489] env[62385]: DEBUG nova.compute.manager [req-e16c58c1-b687-48c7-9338-0820696c5f69 req-2dd4dd66-076b-488e-a02d-2aa3aafb17ff service nova] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Received event network-changed-c308701b-9b9e-4ff3-9a01-c27aabf54f0d {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1030.968662] env[62385]: DEBUG nova.compute.manager [req-e16c58c1-b687-48c7-9338-0820696c5f69 req-2dd4dd66-076b-488e-a02d-2aa3aafb17ff service nova] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Refreshing instance network info cache due to event network-changed-c308701b-9b9e-4ff3-9a01-c27aabf54f0d. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1030.968870] env[62385]: DEBUG oslo_concurrency.lockutils [req-e16c58c1-b687-48c7-9338-0820696c5f69 req-2dd4dd66-076b-488e-a02d-2aa3aafb17ff service nova] Acquiring lock "refresh_cache-68748d25-017f-4b58-9984-70264abbc5ba" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.969017] env[62385]: DEBUG oslo_concurrency.lockutils [req-e16c58c1-b687-48c7-9338-0820696c5f69 req-2dd4dd66-076b-488e-a02d-2aa3aafb17ff service nova] Acquired lock "refresh_cache-68748d25-017f-4b58-9984-70264abbc5ba" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.969194] env[62385]: DEBUG nova.network.neutron [req-e16c58c1-b687-48c7-9338-0820696c5f69 req-2dd4dd66-076b-488e-a02d-2aa3aafb17ff service nova] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Refreshing network info cache for port c308701b-9b9e-4ff3-9a01-c27aabf54f0d {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1030.977640] env[62385]: DEBUG oslo_vmware.api [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Waiting for the task: (returnval){ [ 1030.977640] env[62385]: value = "task-1206381" [ 1030.977640] env[62385]: _type = "Task" [ 1030.977640] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.988886] env[62385]: DEBUG oslo_vmware.api [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Task: {'id': task-1206381, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.415189] env[62385]: INFO nova.compute.manager [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Took 21.98 seconds to build instance. [ 1031.489340] env[62385]: DEBUG nova.compute.manager [req-98a840f5-0488-4448-b926-4e7389312dcb req-5481f2e1-3b0c-4dba-aff9-7723b78fffc9 service nova] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Received event network-vif-plugged-deafdeb9-f59c-48f4-873b-6e4b27d4a688 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1031.490183] env[62385]: DEBUG oslo_concurrency.lockutils [req-98a840f5-0488-4448-b926-4e7389312dcb req-5481f2e1-3b0c-4dba-aff9-7723b78fffc9 service nova] Acquiring lock "8e1d416c-2624-468c-94d7-6265e9f4178f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.490396] env[62385]: DEBUG oslo_concurrency.lockutils [req-98a840f5-0488-4448-b926-4e7389312dcb req-5481f2e1-3b0c-4dba-aff9-7723b78fffc9 service nova] Lock "8e1d416c-2624-468c-94d7-6265e9f4178f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.490575] env[62385]: DEBUG oslo_concurrency.lockutils [req-98a840f5-0488-4448-b926-4e7389312dcb req-5481f2e1-3b0c-4dba-aff9-7723b78fffc9 service nova] Lock "8e1d416c-2624-468c-94d7-6265e9f4178f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.490749] env[62385]: DEBUG nova.compute.manager [req-98a840f5-0488-4448-b926-4e7389312dcb req-5481f2e1-3b0c-4dba-aff9-7723b78fffc9 service nova] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] No waiting events found dispatching network-vif-plugged-deafdeb9-f59c-48f4-873b-6e4b27d4a688 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1031.490918] env[62385]: WARNING nova.compute.manager [req-98a840f5-0488-4448-b926-4e7389312dcb req-5481f2e1-3b0c-4dba-aff9-7723b78fffc9 service nova] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Received unexpected event network-vif-plugged-deafdeb9-f59c-48f4-873b-6e4b27d4a688 for instance with vm_state shelved_offloaded and task_state spawning. [ 1031.497100] env[62385]: DEBUG oslo_vmware.api [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Task: {'id': task-1206381, 'name': ReconfigVM_Task, 'duration_secs': 0.326896} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.499549] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Reconfigured VM instance instance-0000006a to attach disk [datastore2] volume-e44d7ff8-cd5f-4a36-a788-a75683085f24/volume-e44d7ff8-cd5f-4a36-a788-a75683085f24.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1031.505312] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-34e92c67-666a-4cac-81ca-7000d2943f42 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.516741] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f108652f-af52-42ea-89f4-63d61dc8d69d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.520946] env[62385]: DEBUG oslo_vmware.api [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Waiting for the task: (returnval){ [ 1031.520946] env[62385]: value = "task-1206382" [ 1031.520946] env[62385]: _type = "Task" [ 1031.520946] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.527891] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40ac8e1b-8085-478c-b38e-cf3a1df2a389 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.535559] env[62385]: DEBUG oslo_vmware.api [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Task: {'id': task-1206382, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.559600] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c86c42e-0d9a-4537-bfed-0c05e5f8627d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.567257] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-913449a9-e0ee-4479-b0b7-d152374b6536 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.580812] env[62385]: DEBUG nova.compute.provider_tree [None req-c6808a79-c5d5-48a7-9431-20da78d448b1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1031.668485] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquiring lock "refresh_cache-8e1d416c-2624-468c-94d7-6265e9f4178f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.668676] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquired lock "refresh_cache-8e1d416c-2624-468c-94d7-6265e9f4178f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.668860] env[62385]: DEBUG nova.network.neutron [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1031.746009] env[62385]: DEBUG nova.network.neutron [req-e16c58c1-b687-48c7-9338-0820696c5f69 req-2dd4dd66-076b-488e-a02d-2aa3aafb17ff service nova] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Updated VIF entry in instance network info cache for port c308701b-9b9e-4ff3-9a01-c27aabf54f0d. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1031.746403] env[62385]: DEBUG nova.network.neutron [req-e16c58c1-b687-48c7-9338-0820696c5f69 req-2dd4dd66-076b-488e-a02d-2aa3aafb17ff service nova] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Updating instance_info_cache with network_info: [{"id": "c308701b-9b9e-4ff3-9a01-c27aabf54f0d", "address": "fa:16:3e:9a:ed:8c", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc308701b-9b", "ovs_interfaceid": "c308701b-9b9e-4ff3-9a01-c27aabf54f0d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.917352] env[62385]: DEBUG oslo_concurrency.lockutils [None req-51cab2df-9bd7-436b-934a-e23f58f360a7 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "47311a9c-74d2-400d-a22d-9ef55b14edb5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.493s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.030554] env[62385]: DEBUG oslo_vmware.api [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Task: {'id': task-1206382, 'name': ReconfigVM_Task, 'duration_secs': 0.153806} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.030870] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261288', 'volume_id': 'e44d7ff8-cd5f-4a36-a788-a75683085f24', 'name': 'volume-e44d7ff8-cd5f-4a36-a788-a75683085f24', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5a5bcb7c-86cd-4950-a3dd-15515ae0aca6', 'attached_at': '', 'detached_at': '', 'volume_id': 'e44d7ff8-cd5f-4a36-a788-a75683085f24', 'serial': 'e44d7ff8-cd5f-4a36-a788-a75683085f24'} {{(pid=62385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1032.031513] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-93dd92d5-67a4-44c4-80b5-9398e0f92ee4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.038663] env[62385]: DEBUG oslo_vmware.api [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Waiting for the task: (returnval){ [ 1032.038663] env[62385]: value = "task-1206384" [ 1032.038663] env[62385]: _type = "Task" [ 1032.038663] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.051571] env[62385]: DEBUG oslo_vmware.api [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Task: {'id': task-1206384, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.085964] env[62385]: DEBUG nova.scheduler.client.report [None req-c6808a79-c5d5-48a7-9431-20da78d448b1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1032.248872] env[62385]: DEBUG oslo_concurrency.lockutils [req-e16c58c1-b687-48c7-9338-0820696c5f69 req-2dd4dd66-076b-488e-a02d-2aa3aafb17ff service nova] Releasing lock "refresh_cache-68748d25-017f-4b58-9984-70264abbc5ba" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.249132] env[62385]: DEBUG nova.compute.manager [req-e16c58c1-b687-48c7-9338-0820696c5f69 req-2dd4dd66-076b-488e-a02d-2aa3aafb17ff service nova] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Received event network-changed-a7153a2d-db69-4d14-8ddc-b634ddabbc18 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1032.249391] env[62385]: DEBUG nova.compute.manager [req-e16c58c1-b687-48c7-9338-0820696c5f69 req-2dd4dd66-076b-488e-a02d-2aa3aafb17ff service nova] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Refreshing instance network info cache due to event network-changed-a7153a2d-db69-4d14-8ddc-b634ddabbc18. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1032.249546] env[62385]: DEBUG oslo_concurrency.lockutils [req-e16c58c1-b687-48c7-9338-0820696c5f69 req-2dd4dd66-076b-488e-a02d-2aa3aafb17ff service nova] Acquiring lock "refresh_cache-57637d6e-39e2-405c-bbcf-64bf0ffbebd1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.249687] env[62385]: DEBUG oslo_concurrency.lockutils [req-e16c58c1-b687-48c7-9338-0820696c5f69 req-2dd4dd66-076b-488e-a02d-2aa3aafb17ff service nova] Acquired lock "refresh_cache-57637d6e-39e2-405c-bbcf-64bf0ffbebd1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.249844] env[62385]: DEBUG nova.network.neutron [req-e16c58c1-b687-48c7-9338-0820696c5f69 req-2dd4dd66-076b-488e-a02d-2aa3aafb17ff service nova] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Refreshing network info cache for port a7153a2d-db69-4d14-8ddc-b634ddabbc18 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1032.508147] env[62385]: DEBUG nova.network.neutron [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Updating instance_info_cache with network_info: [{"id": "deafdeb9-f59c-48f4-873b-6e4b27d4a688", "address": "fa:16:3e:81:4d:75", "network": {"id": "9c4d6464-cacc-44f8-b7e5-0699cd312c53", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1889153615-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.186", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09c2c5e938bf46fd9353e47aecc72478", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4d3f69a-b086-4c3b-b976-5a848b63dfc4", "external-id": "nsx-vlan-transportzone-627", "segmentation_id": 627, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdeafdeb9-f5", "ovs_interfaceid": "deafdeb9-f59c-48f4-873b-6e4b27d4a688", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1032.552530] env[62385]: DEBUG oslo_vmware.api [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Task: {'id': task-1206384, 'name': Rename_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.591568] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c6808a79-c5d5-48a7-9431-20da78d448b1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.760s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.594132] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c02c14cc-903b-4279-9294-6fa98f845aa2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.327s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.594257] env[62385]: DEBUG nova.objects.instance [None req-c02c14cc-903b-4279-9294-6fa98f845aa2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lazy-loading 'resources' on Instance uuid 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1032.613874] env[62385]: INFO nova.scheduler.client.report [None req-c6808a79-c5d5-48a7-9431-20da78d448b1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Deleted allocations for instance b0a6a3b9-05e4-4190-9cc6-4237dee07f91 [ 1032.943902] env[62385]: DEBUG nova.network.neutron [req-e16c58c1-b687-48c7-9338-0820696c5f69 req-2dd4dd66-076b-488e-a02d-2aa3aafb17ff service nova] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Updated VIF entry in instance network info cache for port a7153a2d-db69-4d14-8ddc-b634ddabbc18. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1032.944291] env[62385]: DEBUG nova.network.neutron [req-e16c58c1-b687-48c7-9338-0820696c5f69 req-2dd4dd66-076b-488e-a02d-2aa3aafb17ff service nova] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Updating instance_info_cache with network_info: [{"id": "a7153a2d-db69-4d14-8ddc-b634ddabbc18", "address": "fa:16:3e:a3:0d:ab", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.210", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7153a2d-db", "ovs_interfaceid": "a7153a2d-db69-4d14-8ddc-b634ddabbc18", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1033.011015] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Releasing lock "refresh_cache-8e1d416c-2624-468c-94d7-6265e9f4178f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.039495] env[62385]: DEBUG nova.virt.hardware [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='988b2c215c83cfc8989e3258bb000435',container_format='bare',created_at=2024-10-09T02:25:36Z,direct_url=,disk_format='vmdk',id=a561779b-d9f7-4373-8a4d-586fbc25c6d7,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1671098717-shelved',owner='09c2c5e938bf46fd9353e47aecc72478',properties=ImageMetaProps,protected=,size=31665664,status='active',tags=,updated_at=2024-10-09T02:25:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1033.039763] env[62385]: DEBUG nova.virt.hardware [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1033.039929] env[62385]: DEBUG nova.virt.hardware [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1033.040152] env[62385]: DEBUG nova.virt.hardware [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1033.040333] env[62385]: DEBUG nova.virt.hardware [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1033.040493] env[62385]: DEBUG nova.virt.hardware [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1033.040720] env[62385]: DEBUG nova.virt.hardware [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1033.040880] env[62385]: DEBUG nova.virt.hardware [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1033.041069] env[62385]: DEBUG nova.virt.hardware [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1033.041248] env[62385]: DEBUG nova.virt.hardware [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1033.041428] env[62385]: DEBUG nova.virt.hardware [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1033.042368] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daec50c5-1bcb-411b-9e46-7d08cd658707 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.055079] env[62385]: DEBUG oslo_vmware.api [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Task: {'id': task-1206384, 'name': Rename_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.057373] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5633ce38-8e82-46c2-9f0a-69ddf90d2b1f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.072237] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:81:4d:75', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c4d3f69a-b086-4c3b-b976-5a848b63dfc4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'deafdeb9-f59c-48f4-873b-6e4b27d4a688', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1033.079683] env[62385]: DEBUG oslo.service.loopingcall [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1033.079940] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1033.080182] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ca9da2eb-75ce-4552-b7fd-5f9d6c674635 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.100698] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1033.100698] env[62385]: value = "task-1206385" [ 1033.100698] env[62385]: _type = "Task" [ 1033.100698] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.108020] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206385, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.120839] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5846f1c4-9189-492c-a7e3-a7c6e834faa4 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "47311a9c-74d2-400d-a22d-9ef55b14edb5" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.121100] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5846f1c4-9189-492c-a7e3-a7c6e834faa4 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "47311a9c-74d2-400d-a22d-9ef55b14edb5" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.121304] env[62385]: DEBUG nova.compute.manager [None req-5846f1c4-9189-492c-a7e3-a7c6e834faa4 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1033.124223] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d3ba569-7811-4a32-aa37-8492794e3738 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.129423] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c6808a79-c5d5-48a7-9431-20da78d448b1 tempest-ServerDiskConfigTestJSON-1635425535 tempest-ServerDiskConfigTestJSON-1635425535-project-member] Lock "b0a6a3b9-05e4-4190-9cc6-4237dee07f91" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.747s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.134159] env[62385]: DEBUG nova.compute.manager [None req-5846f1c4-9189-492c-a7e3-a7c6e834faa4 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62385) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1033.135071] env[62385]: DEBUG nova.objects.instance [None req-5846f1c4-9189-492c-a7e3-a7c6e834faa4 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lazy-loading 'flavor' on Instance uuid 47311a9c-74d2-400d-a22d-9ef55b14edb5 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1033.244507] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fa6e1d1-a66b-44f2-a894-48667dff6de2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.252466] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-828920d2-8800-48bb-a030-53c5328b42f1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.283490] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8358fa65-64a8-461b-a2a4-645747d34fdf {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.291888] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ad0864f-52f5-40a8-8a2a-5a824890c209 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.305879] env[62385]: DEBUG nova.compute.provider_tree [None req-c02c14cc-903b-4279-9294-6fa98f845aa2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1033.447077] env[62385]: DEBUG oslo_concurrency.lockutils [req-e16c58c1-b687-48c7-9338-0820696c5f69 req-2dd4dd66-076b-488e-a02d-2aa3aafb17ff service nova] Releasing lock "refresh_cache-57637d6e-39e2-405c-bbcf-64bf0ffbebd1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.515602] env[62385]: DEBUG nova.compute.manager [req-8c218918-41c7-4f13-a7d2-d8bd3e3768bb req-fd22dfb7-d0d2-4135-b79d-233d36a6a246 service nova] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Received event network-changed-deafdeb9-f59c-48f4-873b-6e4b27d4a688 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1033.516035] env[62385]: DEBUG nova.compute.manager [req-8c218918-41c7-4f13-a7d2-d8bd3e3768bb req-fd22dfb7-d0d2-4135-b79d-233d36a6a246 service nova] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Refreshing instance network info cache due to event network-changed-deafdeb9-f59c-48f4-873b-6e4b27d4a688. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1033.516636] env[62385]: DEBUG oslo_concurrency.lockutils [req-8c218918-41c7-4f13-a7d2-d8bd3e3768bb req-fd22dfb7-d0d2-4135-b79d-233d36a6a246 service nova] Acquiring lock "refresh_cache-8e1d416c-2624-468c-94d7-6265e9f4178f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.516636] env[62385]: DEBUG oslo_concurrency.lockutils [req-8c218918-41c7-4f13-a7d2-d8bd3e3768bb req-fd22dfb7-d0d2-4135-b79d-233d36a6a246 service nova] Acquired lock "refresh_cache-8e1d416c-2624-468c-94d7-6265e9f4178f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.516761] env[62385]: DEBUG nova.network.neutron [req-8c218918-41c7-4f13-a7d2-d8bd3e3768bb req-fd22dfb7-d0d2-4135-b79d-233d36a6a246 service nova] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Refreshing network info cache for port deafdeb9-f59c-48f4-873b-6e4b27d4a688 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1033.555257] env[62385]: DEBUG oslo_vmware.api [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Task: {'id': task-1206384, 'name': Rename_Task, 'duration_secs': 1.16726} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.555567] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1033.555939] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a1df2ceb-8ad7-483f-99a5-3bdf04ace214 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.563077] env[62385]: DEBUG oslo_vmware.api [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Waiting for the task: (returnval){ [ 1033.563077] env[62385]: value = "task-1206386" [ 1033.563077] env[62385]: _type = "Task" [ 1033.563077] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.572583] env[62385]: DEBUG oslo_vmware.api [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Task: {'id': task-1206386, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.610374] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206385, 'name': CreateVM_Task, 'duration_secs': 0.344237} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.610763] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1033.611472] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a561779b-d9f7-4373-8a4d-586fbc25c6d7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.611661] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a561779b-d9f7-4373-8a4d-586fbc25c6d7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.612096] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a561779b-d9f7-4373-8a4d-586fbc25c6d7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1033.612372] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c17fef19-3e46-4a95-8d99-93a73a5f0480 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.617536] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1033.617536] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5216e2de-0ca1-3737-3c7b-413664ae436f" [ 1033.617536] env[62385]: _type = "Task" [ 1033.617536] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.625935] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5216e2de-0ca1-3737-3c7b-413664ae436f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.639946] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5846f1c4-9189-492c-a7e3-a7c6e834faa4 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1033.640237] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cdeb3861-3088-41d7-b9db-22b5f0ab39de {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.647203] env[62385]: DEBUG oslo_vmware.api [None req-5846f1c4-9189-492c-a7e3-a7c6e834faa4 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 1033.647203] env[62385]: value = "task-1206387" [ 1033.647203] env[62385]: _type = "Task" [ 1033.647203] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.657022] env[62385]: DEBUG oslo_vmware.api [None req-5846f1c4-9189-492c-a7e3-a7c6e834faa4 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206387, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.809186] env[62385]: DEBUG nova.scheduler.client.report [None req-c02c14cc-903b-4279-9294-6fa98f845aa2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1034.073185] env[62385]: DEBUG oslo_vmware.api [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Task: {'id': task-1206386, 'name': PowerOnVM_Task, 'duration_secs': 0.506596} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.073478] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1034.073700] env[62385]: INFO nova.compute.manager [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Took 6.63 seconds to spawn the instance on the hypervisor. [ 1034.073889] env[62385]: DEBUG nova.compute.manager [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1034.074668] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fdeb376-c8cb-447e-848b-d90465df34c0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.124261] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-782c0473-e437-423d-bdca-2942fb468e6d tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Volume attach. Driver type: vmdk {{(pid=62385) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1034.124508] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-782c0473-e437-423d-bdca-2942fb468e6d tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261298', 'volume_id': '67e47839-a66b-4032-b1c4-93171c686c2f', 'name': 'volume-67e47839-a66b-4032-b1c4-93171c686c2f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a1c71d8b-0953-4cc2-961d-0c2b2fc11110', 'attached_at': '', 'detached_at': '', 'volume_id': '67e47839-a66b-4032-b1c4-93171c686c2f', 'serial': '67e47839-a66b-4032-b1c4-93171c686c2f'} {{(pid=62385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1034.125343] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9e8b9b6-feca-4c75-9e6f-0e51da56d1f0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.131913] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a561779b-d9f7-4373-8a4d-586fbc25c6d7" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.132168] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Processing image a561779b-d9f7-4373-8a4d-586fbc25c6d7 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1034.132429] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a561779b-d9f7-4373-8a4d-586fbc25c6d7/a561779b-d9f7-4373-8a4d-586fbc25c6d7.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.132596] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a561779b-d9f7-4373-8a4d-586fbc25c6d7/a561779b-d9f7-4373-8a4d-586fbc25c6d7.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.132828] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1034.133024] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-675ff2ec-4331-4450-b3b7-4334c599f966 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.148223] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e369eb1a-9237-4ede-aede-720e9fade722 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.155483] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1034.155744] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1034.174949] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8b17064c-4fc6-475e-9529-fe0e661adba3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.185618] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-782c0473-e437-423d-bdca-2942fb468e6d tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] volume-67e47839-a66b-4032-b1c4-93171c686c2f/volume-67e47839-a66b-4032-b1c4-93171c686c2f.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1034.187660] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-17bf6eca-7765-467e-9886-ebafe91ca63f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.203794] env[62385]: DEBUG oslo_vmware.api [None req-5846f1c4-9189-492c-a7e3-a7c6e834faa4 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206387, 'name': PowerOffVM_Task, 'duration_secs': 0.232831} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.206990] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5846f1c4-9189-492c-a7e3-a7c6e834faa4 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1034.207191] env[62385]: DEBUG nova.compute.manager [None req-5846f1c4-9189-492c-a7e3-a7c6e834faa4 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1034.208521] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0931194b-3503-4659-81aa-d110acb6bf68 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.216093] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1034.216093] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]529c84e5-6e56-5fed-f29b-4ca9dc6cc803" [ 1034.216093] env[62385]: _type = "Task" [ 1034.216093] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.216093] env[62385]: DEBUG oslo_vmware.api [None req-782c0473-e437-423d-bdca-2942fb468e6d tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for the task: (returnval){ [ 1034.216093] env[62385]: value = "task-1206388" [ 1034.216093] env[62385]: _type = "Task" [ 1034.216093] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.236178] env[62385]: DEBUG oslo_vmware.api [None req-782c0473-e437-423d-bdca-2942fb468e6d tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206388, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.236746] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Preparing fetch location {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1034.237113] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Fetch image to [datastore1] OSTACK_IMG_dc01d2d7-f982-46ac-bca5-1b922f01ffe5/OSTACK_IMG_dc01d2d7-f982-46ac-bca5-1b922f01ffe5.vmdk {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1034.237519] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Downloading stream optimized image a561779b-d9f7-4373-8a4d-586fbc25c6d7 to [datastore1] OSTACK_IMG_dc01d2d7-f982-46ac-bca5-1b922f01ffe5/OSTACK_IMG_dc01d2d7-f982-46ac-bca5-1b922f01ffe5.vmdk on the data store datastore1 as vApp {{(pid=62385) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1034.237922] env[62385]: DEBUG nova.virt.vmwareapi.images [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Downloading image file data a561779b-d9f7-4373-8a4d-586fbc25c6d7 to the ESX as VM named 'OSTACK_IMG_dc01d2d7-f982-46ac-bca5-1b922f01ffe5' {{(pid=62385) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1034.314186] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c02c14cc-903b-4279-9294-6fa98f845aa2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.720s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.316919] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.900s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.317900] env[62385]: INFO nova.compute.claims [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1034.333708] env[62385]: INFO nova.scheduler.client.report [None req-c02c14cc-903b-4279-9294-6fa98f845aa2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Deleted allocations for instance 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e [ 1034.343546] env[62385]: DEBUG oslo_vmware.rw_handles [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1034.343546] env[62385]: value = "resgroup-9" [ 1034.343546] env[62385]: _type = "ResourcePool" [ 1034.343546] env[62385]: }. {{(pid=62385) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1034.343826] env[62385]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-37bd851c-2152-43b6-8d52-83f95e0522a7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.368463] env[62385]: DEBUG oslo_vmware.rw_handles [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lease: (returnval){ [ 1034.368463] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a6c238-d681-84c7-ab47-6964d9e9ca02" [ 1034.368463] env[62385]: _type = "HttpNfcLease" [ 1034.368463] env[62385]: } obtained for vApp import into resource pool (val){ [ 1034.368463] env[62385]: value = "resgroup-9" [ 1034.368463] env[62385]: _type = "ResourcePool" [ 1034.368463] env[62385]: }. {{(pid=62385) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1034.368919] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the lease: (returnval){ [ 1034.368919] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a6c238-d681-84c7-ab47-6964d9e9ca02" [ 1034.368919] env[62385]: _type = "HttpNfcLease" [ 1034.368919] env[62385]: } to be ready. {{(pid=62385) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1034.377380] env[62385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1034.377380] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a6c238-d681-84c7-ab47-6964d9e9ca02" [ 1034.377380] env[62385]: _type = "HttpNfcLease" [ 1034.377380] env[62385]: } is initializing. {{(pid=62385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1034.510983] env[62385]: DEBUG nova.network.neutron [req-8c218918-41c7-4f13-a7d2-d8bd3e3768bb req-fd22dfb7-d0d2-4135-b79d-233d36a6a246 service nova] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Updated VIF entry in instance network info cache for port deafdeb9-f59c-48f4-873b-6e4b27d4a688. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1034.511430] env[62385]: DEBUG nova.network.neutron [req-8c218918-41c7-4f13-a7d2-d8bd3e3768bb req-fd22dfb7-d0d2-4135-b79d-233d36a6a246 service nova] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Updating instance_info_cache with network_info: [{"id": "deafdeb9-f59c-48f4-873b-6e4b27d4a688", "address": "fa:16:3e:81:4d:75", "network": {"id": "9c4d6464-cacc-44f8-b7e5-0699cd312c53", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1889153615-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.186", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09c2c5e938bf46fd9353e47aecc72478", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4d3f69a-b086-4c3b-b976-5a848b63dfc4", "external-id": "nsx-vlan-transportzone-627", "segmentation_id": 627, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdeafdeb9-f5", "ovs_interfaceid": "deafdeb9-f59c-48f4-873b-6e4b27d4a688", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.596984] env[62385]: INFO nova.compute.manager [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Took 19.93 seconds to build instance. [ 1034.728822] env[62385]: DEBUG oslo_vmware.api [None req-782c0473-e437-423d-bdca-2942fb468e6d tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206388, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.729321] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5846f1c4-9189-492c-a7e3-a7c6e834faa4 tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "47311a9c-74d2-400d-a22d-9ef55b14edb5" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.608s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.844134] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c02c14cc-903b-4279-9294-6fa98f845aa2 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "9b9fd167-d843-4b9d-9e4f-dbde5d3d629e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.615s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.877889] env[62385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1034.877889] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a6c238-d681-84c7-ab47-6964d9e9ca02" [ 1034.877889] env[62385]: _type = "HttpNfcLease" [ 1034.877889] env[62385]: } is initializing. {{(pid=62385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1035.014328] env[62385]: DEBUG oslo_concurrency.lockutils [req-8c218918-41c7-4f13-a7d2-d8bd3e3768bb req-fd22dfb7-d0d2-4135-b79d-233d36a6a246 service nova] Releasing lock "refresh_cache-8e1d416c-2624-468c-94d7-6265e9f4178f" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.099420] env[62385]: DEBUG oslo_concurrency.lockutils [None req-7fe4b295-836f-4f6a-abf7-691dc54059c1 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Lock "5a5bcb7c-86cd-4950-a3dd-15515ae0aca6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.442s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.227451] env[62385]: DEBUG oslo_vmware.api [None req-782c0473-e437-423d-bdca-2942fb468e6d tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206388, 'name': ReconfigVM_Task, 'duration_secs': 0.89952} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.227754] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-782c0473-e437-423d-bdca-2942fb468e6d tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Reconfigured VM instance instance-00000063 to attach disk [datastore1] volume-67e47839-a66b-4032-b1c4-93171c686c2f/volume-67e47839-a66b-4032-b1c4-93171c686c2f.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1035.232852] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cfc3d162-69f7-4a20-9b1f-2a34a98bfd59 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.254147] env[62385]: DEBUG nova.compute.manager [req-005ba9b2-af80-4999-97ee-7657551b94c7 req-ca42e03a-f226-49d7-910d-a897ddd605cf service nova] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Received event network-changed-5af3050a-7535-42c3-bb8d-4f1141049e32 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1035.255336] env[62385]: DEBUG nova.compute.manager [req-005ba9b2-af80-4999-97ee-7657551b94c7 req-ca42e03a-f226-49d7-910d-a897ddd605cf service nova] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Refreshing instance network info cache due to event network-changed-5af3050a-7535-42c3-bb8d-4f1141049e32. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1035.255336] env[62385]: DEBUG oslo_concurrency.lockutils [req-005ba9b2-af80-4999-97ee-7657551b94c7 req-ca42e03a-f226-49d7-910d-a897ddd605cf service nova] Acquiring lock "refresh_cache-5a5bcb7c-86cd-4950-a3dd-15515ae0aca6" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.255336] env[62385]: DEBUG oslo_concurrency.lockutils [req-005ba9b2-af80-4999-97ee-7657551b94c7 req-ca42e03a-f226-49d7-910d-a897ddd605cf service nova] Acquired lock "refresh_cache-5a5bcb7c-86cd-4950-a3dd-15515ae0aca6" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.255526] env[62385]: DEBUG nova.network.neutron [req-005ba9b2-af80-4999-97ee-7657551b94c7 req-ca42e03a-f226-49d7-910d-a897ddd605cf service nova] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Refreshing network info cache for port 5af3050a-7535-42c3-bb8d-4f1141049e32 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1035.257625] env[62385]: DEBUG oslo_vmware.api [None req-782c0473-e437-423d-bdca-2942fb468e6d tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for the task: (returnval){ [ 1035.257625] env[62385]: value = "task-1206390" [ 1035.257625] env[62385]: _type = "Task" [ 1035.257625] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.269652] env[62385]: DEBUG oslo_vmware.api [None req-782c0473-e437-423d-bdca-2942fb468e6d tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206390, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.380446] env[62385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1035.380446] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a6c238-d681-84c7-ab47-6964d9e9ca02" [ 1035.380446] env[62385]: _type = "HttpNfcLease" [ 1035.380446] env[62385]: } is initializing. {{(pid=62385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1035.476684] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96a0d141-08a3-4ee5-88ff-6dab6fd40333 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.488440] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45d62a42-9269-4f85-ba57-9527e9a22004 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.530109] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26ac1c9d-8b8d-42c0-9562-585300666e0a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.540811] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79b4498d-353e-4408-abf1-7f81e70eff3f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.566446] env[62385]: DEBUG nova.compute.provider_tree [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1035.568465] env[62385]: DEBUG oslo_concurrency.lockutils [None req-821932eb-b83d-456c-bbce-fef50c0c3b8b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "47311a9c-74d2-400d-a22d-9ef55b14edb5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.568710] env[62385]: DEBUG oslo_concurrency.lockutils [None req-821932eb-b83d-456c-bbce-fef50c0c3b8b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "47311a9c-74d2-400d-a22d-9ef55b14edb5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.568924] env[62385]: DEBUG oslo_concurrency.lockutils [None req-821932eb-b83d-456c-bbce-fef50c0c3b8b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "47311a9c-74d2-400d-a22d-9ef55b14edb5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.569188] env[62385]: DEBUG oslo_concurrency.lockutils [None req-821932eb-b83d-456c-bbce-fef50c0c3b8b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "47311a9c-74d2-400d-a22d-9ef55b14edb5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.569448] env[62385]: DEBUG oslo_concurrency.lockutils [None req-821932eb-b83d-456c-bbce-fef50c0c3b8b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "47311a9c-74d2-400d-a22d-9ef55b14edb5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.571834] env[62385]: INFO nova.compute.manager [None req-821932eb-b83d-456c-bbce-fef50c0c3b8b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Terminating instance [ 1035.573520] env[62385]: DEBUG nova.compute.manager [None req-821932eb-b83d-456c-bbce-fef50c0c3b8b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1035.574019] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-821932eb-b83d-456c-bbce-fef50c0c3b8b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1035.574787] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1505b0a-3bae-4224-a0b1-3f06ace2aca0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.583415] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-821932eb-b83d-456c-bbce-fef50c0c3b8b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1035.583415] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4112291e-b7c1-420b-a14e-1317d0de7a1e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.674309] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-821932eb-b83d-456c-bbce-fef50c0c3b8b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1035.674612] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-821932eb-b83d-456c-bbce-fef50c0c3b8b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1035.674612] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-821932eb-b83d-456c-bbce-fef50c0c3b8b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Deleting the datastore file [datastore1] 47311a9c-74d2-400d-a22d-9ef55b14edb5 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1035.674928] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-924c06ba-bd80-476f-ab4c-a9d4de50a3fe {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.683550] env[62385]: DEBUG oslo_vmware.api [None req-821932eb-b83d-456c-bbce-fef50c0c3b8b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for the task: (returnval){ [ 1035.683550] env[62385]: value = "task-1206392" [ 1035.683550] env[62385]: _type = "Task" [ 1035.683550] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.692766] env[62385]: DEBUG oslo_vmware.api [None req-821932eb-b83d-456c-bbce-fef50c0c3b8b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206392, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.770407] env[62385]: DEBUG oslo_vmware.api [None req-782c0473-e437-423d-bdca-2942fb468e6d tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206390, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.878767] env[62385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1035.878767] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a6c238-d681-84c7-ab47-6964d9e9ca02" [ 1035.878767] env[62385]: _type = "HttpNfcLease" [ 1035.878767] env[62385]: } is ready. {{(pid=62385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1035.879565] env[62385]: DEBUG oslo_vmware.rw_handles [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1035.879565] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a6c238-d681-84c7-ab47-6964d9e9ca02" [ 1035.879565] env[62385]: _type = "HttpNfcLease" [ 1035.879565] env[62385]: }. {{(pid=62385) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1035.879812] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32e8d8df-f4df-4527-aea1-eeae9fbdb8a3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.888398] env[62385]: DEBUG oslo_vmware.rw_handles [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5251f045-e8ae-3c32-a761-0548d67ec0f2/disk-0.vmdk from lease info. {{(pid=62385) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1035.888600] env[62385]: DEBUG oslo_vmware.rw_handles [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Creating HTTP connection to write to file with size = 31665664 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5251f045-e8ae-3c32-a761-0548d67ec0f2/disk-0.vmdk. {{(pid=62385) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1035.964352] env[62385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-2385e10e-dcf3-4314-9ed6-02bc8ac720e5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.073110] env[62385]: DEBUG nova.scheduler.client.report [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1036.133186] env[62385]: DEBUG nova.network.neutron [req-005ba9b2-af80-4999-97ee-7657551b94c7 req-ca42e03a-f226-49d7-910d-a897ddd605cf service nova] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Updated VIF entry in instance network info cache for port 5af3050a-7535-42c3-bb8d-4f1141049e32. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1036.133581] env[62385]: DEBUG nova.network.neutron [req-005ba9b2-af80-4999-97ee-7657551b94c7 req-ca42e03a-f226-49d7-910d-a897ddd605cf service nova] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Updating instance_info_cache with network_info: [{"id": "5af3050a-7535-42c3-bb8d-4f1141049e32", "address": "fa:16:3e:3d:8b:5f", "network": {"id": "b428b89a-516a-4b8b-9a77-8d922c1865aa", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-105102189-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58f5feadcb364e16bf9e23e8c6990318", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "62f28d75-4e6a-4ae5-b8b3-d0652ea26d08", "external-id": "nsx-vlan-transportzone-801", "segmentation_id": 801, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5af3050a-75", "ovs_interfaceid": "5af3050a-7535-42c3-bb8d-4f1141049e32", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.197212] env[62385]: DEBUG oslo_vmware.api [None req-821932eb-b83d-456c-bbce-fef50c0c3b8b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Task: {'id': task-1206392, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.336045} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.197514] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-821932eb-b83d-456c-bbce-fef50c0c3b8b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1036.197721] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-821932eb-b83d-456c-bbce-fef50c0c3b8b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1036.198626] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-821932eb-b83d-456c-bbce-fef50c0c3b8b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1036.198626] env[62385]: INFO nova.compute.manager [None req-821932eb-b83d-456c-bbce-fef50c0c3b8b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Took 0.62 seconds to destroy the instance on the hypervisor. [ 1036.198626] env[62385]: DEBUG oslo.service.loopingcall [None req-821932eb-b83d-456c-bbce-fef50c0c3b8b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1036.198626] env[62385]: DEBUG nova.compute.manager [-] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1036.198858] env[62385]: DEBUG nova.network.neutron [-] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1036.281269] env[62385]: DEBUG oslo_vmware.api [None req-782c0473-e437-423d-bdca-2942fb468e6d tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206390, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.580066] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.263s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.580066] env[62385]: DEBUG nova.compute.manager [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1036.584549] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9a98fbb0-158e-4d6e-95e6-91542f6881e4 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.844s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.584782] env[62385]: DEBUG nova.objects.instance [None req-9a98fbb0-158e-4d6e-95e6-91542f6881e4 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lazy-loading 'resources' on Instance uuid 5daf0a99-0c2a-40d8-afc7-1998e21e32d8 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1036.639755] env[62385]: DEBUG oslo_concurrency.lockutils [req-005ba9b2-af80-4999-97ee-7657551b94c7 req-ca42e03a-f226-49d7-910d-a897ddd605cf service nova] Releasing lock "refresh_cache-5a5bcb7c-86cd-4950-a3dd-15515ae0aca6" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1036.725786] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "1289e63a-5a9d-4615-b6cf-bbf5f4ae0317" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.726067] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "1289e63a-5a9d-4615-b6cf-bbf5f4ae0317" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.751591] env[62385]: DEBUG nova.compute.manager [req-762affbb-0c66-476d-b99b-e7b62c9bccee req-1ee05155-c9e2-426e-b427-92d47a5634f5 service nova] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Received event network-vif-deleted-4af1e8e6-edeb-47ff-b1b1-f4f974a69544 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1036.751801] env[62385]: INFO nova.compute.manager [req-762affbb-0c66-476d-b99b-e7b62c9bccee req-1ee05155-c9e2-426e-b427-92d47a5634f5 service nova] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Neutron deleted interface 4af1e8e6-edeb-47ff-b1b1-f4f974a69544; detaching it from the instance and deleting it from the info cache [ 1036.751980] env[62385]: DEBUG nova.network.neutron [req-762affbb-0c66-476d-b99b-e7b62c9bccee req-1ee05155-c9e2-426e-b427-92d47a5634f5 service nova] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.781271] env[62385]: DEBUG oslo_vmware.api [None req-782c0473-e437-423d-bdca-2942fb468e6d tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206390, 'name': ReconfigVM_Task, 'duration_secs': 1.204922} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.783550] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-782c0473-e437-423d-bdca-2942fb468e6d tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261298', 'volume_id': '67e47839-a66b-4032-b1c4-93171c686c2f', 'name': 'volume-67e47839-a66b-4032-b1c4-93171c686c2f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a1c71d8b-0953-4cc2-961d-0c2b2fc11110', 'attached_at': '', 'detached_at': '', 'volume_id': '67e47839-a66b-4032-b1c4-93171c686c2f', 'serial': '67e47839-a66b-4032-b1c4-93171c686c2f'} {{(pid=62385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1037.087344] env[62385]: DEBUG nova.compute.utils [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1037.094083] env[62385]: DEBUG nova.compute.manager [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1037.094291] env[62385]: DEBUG nova.network.neutron [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1037.176424] env[62385]: DEBUG nova.policy [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b178ea9a750645bea32dcc2873304ec0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5d716d19f79945db98f684df0cf4b302', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 1037.205115] env[62385]: DEBUG oslo_vmware.rw_handles [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Completed reading data from the image iterator. {{(pid=62385) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1037.205389] env[62385]: DEBUG oslo_vmware.rw_handles [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5251f045-e8ae-3c32-a761-0548d67ec0f2/disk-0.vmdk. {{(pid=62385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1037.207023] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e84964ad-de44-4f79-8b68-be7d6dd6c579 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.220696] env[62385]: DEBUG oslo_vmware.rw_handles [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5251f045-e8ae-3c32-a761-0548d67ec0f2/disk-0.vmdk is in state: ready. {{(pid=62385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1037.221254] env[62385]: DEBUG oslo_vmware.rw_handles [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5251f045-e8ae-3c32-a761-0548d67ec0f2/disk-0.vmdk. {{(pid=62385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1037.221717] env[62385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-f27eb2c3-fff2-4fdf-b1cc-8448a0c52497 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.227904] env[62385]: DEBUG nova.network.neutron [-] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1037.229930] env[62385]: DEBUG nova.compute.manager [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1037.257436] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-40769c8d-a46f-4734-9bd7-887a1741fa9f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.268225] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4ece468-14e6-4d79-9764-5d4218394c8c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.276761] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e0806ec-49f1-4a99-9562-ca720f0c7851 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.298978] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb30d22c-0dca-4e77-b733-e92075c942e9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.342863] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-642bbe7d-3c43-4030-a739-0d4e218d858b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.346298] env[62385]: DEBUG nova.compute.manager [req-762affbb-0c66-476d-b99b-e7b62c9bccee req-1ee05155-c9e2-426e-b427-92d47a5634f5 service nova] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Detach interface failed, port_id=4af1e8e6-edeb-47ff-b1b1-f4f974a69544, reason: Instance 47311a9c-74d2-400d-a22d-9ef55b14edb5 could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1037.352921] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69fb3f24-5734-4dc1-af2d-b48e43ec127a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.370546] env[62385]: DEBUG nova.compute.provider_tree [None req-9a98fbb0-158e-4d6e-95e6-91542f6881e4 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1037.576256] env[62385]: DEBUG oslo_vmware.rw_handles [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5251f045-e8ae-3c32-a761-0548d67ec0f2/disk-0.vmdk. {{(pid=62385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1037.576631] env[62385]: INFO nova.virt.vmwareapi.images [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Downloaded image file data a561779b-d9f7-4373-8a4d-586fbc25c6d7 [ 1037.577830] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11a4ecb7-ea96-404a-84da-27deb4a2faa5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.598942] env[62385]: DEBUG nova.compute.manager [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1037.602678] env[62385]: DEBUG nova.network.neutron [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Successfully created port: b19d5597-14de-4b08-b628-3bc8e539f77a {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1037.604271] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8e9b7499-4064-4981-b9e9-84b1b8e40147 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.736625] env[62385]: INFO nova.compute.manager [-] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Took 1.54 seconds to deallocate network for instance. [ 1037.758397] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.826264] env[62385]: INFO nova.virt.vmwareapi.images [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] The imported VM was unregistered [ 1037.832020] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Caching image {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1037.832020] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Creating directory with path [datastore1] devstack-image-cache_base/a561779b-d9f7-4373-8a4d-586fbc25c6d7 {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1037.832020] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-929e0a90-37bc-44fe-a556-ab88b6d35899 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.850142] env[62385]: DEBUG nova.objects.instance [None req-782c0473-e437-423d-bdca-2942fb468e6d tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lazy-loading 'flavor' on Instance uuid a1c71d8b-0953-4cc2-961d-0c2b2fc11110 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1037.873949] env[62385]: DEBUG nova.scheduler.client.report [None req-9a98fbb0-158e-4d6e-95e6-91542f6881e4 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1037.878910] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Created directory with path [datastore1] devstack-image-cache_base/a561779b-d9f7-4373-8a4d-586fbc25c6d7 {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1037.879039] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_dc01d2d7-f982-46ac-bca5-1b922f01ffe5/OSTACK_IMG_dc01d2d7-f982-46ac-bca5-1b922f01ffe5.vmdk to [datastore1] devstack-image-cache_base/a561779b-d9f7-4373-8a4d-586fbc25c6d7/a561779b-d9f7-4373-8a4d-586fbc25c6d7.vmdk. {{(pid=62385) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1037.879396] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-a2d2e46f-b0aa-48ea-8d79-ff0d516d44c2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.887014] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1037.887014] env[62385]: value = "task-1206394" [ 1037.887014] env[62385]: _type = "Task" [ 1037.887014] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.895194] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206394, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.214738] env[62385]: DEBUG oslo_concurrency.lockutils [None req-56e99cef-0ff6-4dac-b865-2525195eae09 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Acquiring lock "a1c71d8b-0953-4cc2-961d-0c2b2fc11110" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.248608] env[62385]: DEBUG oslo_concurrency.lockutils [None req-821932eb-b83d-456c-bbce-fef50c0c3b8b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.357918] env[62385]: DEBUG oslo_concurrency.lockutils [None req-782c0473-e437-423d-bdca-2942fb468e6d tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "a1c71d8b-0953-4cc2-961d-0c2b2fc11110" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.839s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.359272] env[62385]: DEBUG oslo_concurrency.lockutils [None req-56e99cef-0ff6-4dac-b865-2525195eae09 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "a1c71d8b-0953-4cc2-961d-0c2b2fc11110" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.145s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.381060] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9a98fbb0-158e-4d6e-95e6-91542f6881e4 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.796s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.384477] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.626s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.386709] env[62385]: INFO nova.compute.claims [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1038.404358] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206394, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.405660] env[62385]: INFO nova.scheduler.client.report [None req-9a98fbb0-158e-4d6e-95e6-91542f6881e4 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Deleted allocations for instance 5daf0a99-0c2a-40d8-afc7-1998e21e32d8 [ 1038.613017] env[62385]: DEBUG nova.compute.manager [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1038.638248] env[62385]: DEBUG nova.virt.hardware [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1038.638502] env[62385]: DEBUG nova.virt.hardware [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1038.638659] env[62385]: DEBUG nova.virt.hardware [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1038.638839] env[62385]: DEBUG nova.virt.hardware [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1038.639031] env[62385]: DEBUG nova.virt.hardware [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1038.639235] env[62385]: DEBUG nova.virt.hardware [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1038.639462] env[62385]: DEBUG nova.virt.hardware [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1038.639644] env[62385]: DEBUG nova.virt.hardware [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1038.639795] env[62385]: DEBUG nova.virt.hardware [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1038.639963] env[62385]: DEBUG nova.virt.hardware [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1038.640169] env[62385]: DEBUG nova.virt.hardware [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1038.641091] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bf05846-a8ca-4463-9d83-9c368b1d3886 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.649365] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-821f723a-1332-4202-861a-0f72a4004770 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.865340] env[62385]: INFO nova.compute.manager [None req-56e99cef-0ff6-4dac-b865-2525195eae09 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Detaching volume 3b02b9da-5324-4812-abd7-09ff15288694 [ 1038.902908] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206394, 'name': MoveVirtualDisk_Task} progress is 18%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.904380] env[62385]: INFO nova.virt.block_device [None req-56e99cef-0ff6-4dac-b865-2525195eae09 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Attempting to driver detach volume 3b02b9da-5324-4812-abd7-09ff15288694 from mountpoint /dev/sdb [ 1038.904606] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-56e99cef-0ff6-4dac-b865-2525195eae09 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Volume detach. Driver type: vmdk {{(pid=62385) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1038.904799] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-56e99cef-0ff6-4dac-b865-2525195eae09 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261293', 'volume_id': '3b02b9da-5324-4812-abd7-09ff15288694', 'name': 'volume-3b02b9da-5324-4812-abd7-09ff15288694', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a1c71d8b-0953-4cc2-961d-0c2b2fc11110', 'attached_at': '', 'detached_at': '', 'volume_id': '3b02b9da-5324-4812-abd7-09ff15288694', 'serial': '3b02b9da-5324-4812-abd7-09ff15288694'} {{(pid=62385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1038.905681] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae1f6dda-2572-42c4-b3e5-fd1457636fa3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.933081] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9a98fbb0-158e-4d6e-95e6-91542f6881e4 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "5daf0a99-0c2a-40d8-afc7-1998e21e32d8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.538s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.934768] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8030d603-15f0-42f9-8da4-fa9effa8239e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.944737] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86a249a1-b921-4bcd-bf4b-9f43013faa55 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.972191] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-817a28ec-7d07-4932-95b7-f9fa7d6ecef4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.991741] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-56e99cef-0ff6-4dac-b865-2525195eae09 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] The volume has not been displaced from its original location: [datastore1] volume-3b02b9da-5324-4812-abd7-09ff15288694/volume-3b02b9da-5324-4812-abd7-09ff15288694.vmdk. No consolidation needed. {{(pid=62385) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1038.997455] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-56e99cef-0ff6-4dac-b865-2525195eae09 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Reconfiguring VM instance instance-00000063 to detach disk 2001 {{(pid=62385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1038.998140] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1d337ec9-d1ed-4b4a-973b-c535bdc4aea6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.017962] env[62385]: DEBUG oslo_vmware.api [None req-56e99cef-0ff6-4dac-b865-2525195eae09 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for the task: (returnval){ [ 1039.017962] env[62385]: value = "task-1206395" [ 1039.017962] env[62385]: _type = "Task" [ 1039.017962] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.021764] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fde088e5-3eb5-4961-b95b-023de5dbe558 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "interface-57637d6e-39e2-405c-bbcf-64bf0ffbebd1-331b6188-b103-437e-8436-5259bc356e31" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.022036] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fde088e5-3eb5-4961-b95b-023de5dbe558 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "interface-57637d6e-39e2-405c-bbcf-64bf0ffbebd1-331b6188-b103-437e-8436-5259bc356e31" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.022432] env[62385]: DEBUG nova.objects.instance [None req-fde088e5-3eb5-4961-b95b-023de5dbe558 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lazy-loading 'flavor' on Instance uuid 57637d6e-39e2-405c-bbcf-64bf0ffbebd1 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1039.029538] env[62385]: DEBUG oslo_vmware.api [None req-56e99cef-0ff6-4dac-b865-2525195eae09 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206395, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.345884] env[62385]: DEBUG nova.compute.manager [req-c9cc1e63-1634-45f2-8643-6988ba791873 req-89a34081-d01b-424e-a127-b3d46426e89d service nova] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Received event network-vif-plugged-b19d5597-14de-4b08-b628-3bc8e539f77a {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1039.346134] env[62385]: DEBUG oslo_concurrency.lockutils [req-c9cc1e63-1634-45f2-8643-6988ba791873 req-89a34081-d01b-424e-a127-b3d46426e89d service nova] Acquiring lock "c49d4366-2ea0-44f2-8463-daf62bd1c40c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.346353] env[62385]: DEBUG oslo_concurrency.lockutils [req-c9cc1e63-1634-45f2-8643-6988ba791873 req-89a34081-d01b-424e-a127-b3d46426e89d service nova] Lock "c49d4366-2ea0-44f2-8463-daf62bd1c40c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.346532] env[62385]: DEBUG oslo_concurrency.lockutils [req-c9cc1e63-1634-45f2-8643-6988ba791873 req-89a34081-d01b-424e-a127-b3d46426e89d service nova] Lock "c49d4366-2ea0-44f2-8463-daf62bd1c40c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.346703] env[62385]: DEBUG nova.compute.manager [req-c9cc1e63-1634-45f2-8643-6988ba791873 req-89a34081-d01b-424e-a127-b3d46426e89d service nova] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] No waiting events found dispatching network-vif-plugged-b19d5597-14de-4b08-b628-3bc8e539f77a {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1039.346903] env[62385]: WARNING nova.compute.manager [req-c9cc1e63-1634-45f2-8643-6988ba791873 req-89a34081-d01b-424e-a127-b3d46426e89d service nova] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Received unexpected event network-vif-plugged-b19d5597-14de-4b08-b628-3bc8e539f77a for instance with vm_state building and task_state spawning. [ 1039.379435] env[62385]: DEBUG nova.network.neutron [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Successfully updated port: b19d5597-14de-4b08-b628-3bc8e539f77a {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1039.407814] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206394, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.551845] env[62385]: DEBUG oslo_vmware.api [None req-56e99cef-0ff6-4dac-b865-2525195eae09 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206395, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.553787] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8212842e-e81d-4bf5-b9e6-26de3b75582a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.564934] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be9c90c5-ccd3-4a00-86da-66001635bd6a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.605527] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24a7b6a2-9808-4326-a86b-c01ada9c0f22 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.615595] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbe65886-fcdc-4991-a6ff-646719cd1113 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.632672] env[62385]: DEBUG nova.compute.provider_tree [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1039.726532] env[62385]: DEBUG nova.objects.instance [None req-fde088e5-3eb5-4961-b95b-023de5dbe558 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lazy-loading 'pci_requests' on Instance uuid 57637d6e-39e2-405c-bbcf-64bf0ffbebd1 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1039.882349] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Acquiring lock "refresh_cache-c49d4366-2ea0-44f2-8463-daf62bd1c40c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.882564] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Acquired lock "refresh_cache-c49d4366-2ea0-44f2-8463-daf62bd1c40c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.882787] env[62385]: DEBUG nova.network.neutron [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1039.904049] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206394, 'name': MoveVirtualDisk_Task} progress is 60%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.035021] env[62385]: DEBUG oslo_vmware.api [None req-56e99cef-0ff6-4dac-b865-2525195eae09 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206395, 'name': ReconfigVM_Task, 'duration_secs': 0.758136} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.035021] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-56e99cef-0ff6-4dac-b865-2525195eae09 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Reconfigured VM instance instance-00000063 to detach disk 2001 {{(pid=62385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1040.037126] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6b8b5d05-7b84-407d-a229-51fc7ea9fc68 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.054102] env[62385]: DEBUG oslo_vmware.api [None req-56e99cef-0ff6-4dac-b865-2525195eae09 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for the task: (returnval){ [ 1040.054102] env[62385]: value = "task-1206396" [ 1040.054102] env[62385]: _type = "Task" [ 1040.054102] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.063773] env[62385]: DEBUG oslo_vmware.api [None req-56e99cef-0ff6-4dac-b865-2525195eae09 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206396, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.136555] env[62385]: DEBUG nova.scheduler.client.report [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1040.230693] env[62385]: DEBUG nova.objects.base [None req-fde088e5-3eb5-4961-b95b-023de5dbe558 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Object Instance<57637d6e-39e2-405c-bbcf-64bf0ffbebd1> lazy-loaded attributes: flavor,pci_requests {{(pid=62385) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1040.230945] env[62385]: DEBUG nova.network.neutron [None req-fde088e5-3eb5-4961-b95b-023de5dbe558 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1040.339350] env[62385]: DEBUG nova.policy [None req-fde088e5-3eb5-4961-b95b-023de5dbe558 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5040b20c848c4d28a198f1773a3ff21e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '012426f098ce40c3aaa00f628fe9cebb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 1040.407506] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206394, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.446346] env[62385]: DEBUG nova.network.neutron [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1040.565669] env[62385]: DEBUG oslo_vmware.api [None req-56e99cef-0ff6-4dac-b865-2525195eae09 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206396, 'name': ReconfigVM_Task, 'duration_secs': 0.224597} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.565995] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-56e99cef-0ff6-4dac-b865-2525195eae09 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261293', 'volume_id': '3b02b9da-5324-4812-abd7-09ff15288694', 'name': 'volume-3b02b9da-5324-4812-abd7-09ff15288694', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a1c71d8b-0953-4cc2-961d-0c2b2fc11110', 'attached_at': '', 'detached_at': '', 'volume_id': '3b02b9da-5324-4812-abd7-09ff15288694', 'serial': '3b02b9da-5324-4812-abd7-09ff15288694'} {{(pid=62385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1040.641808] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.257s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.642464] env[62385]: DEBUG nova.compute.manager [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1040.645392] env[62385]: DEBUG oslo_concurrency.lockutils [None req-821932eb-b83d-456c-bbce-fef50c0c3b8b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.397s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.645695] env[62385]: DEBUG nova.objects.instance [None req-821932eb-b83d-456c-bbce-fef50c0c3b8b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lazy-loading 'resources' on Instance uuid 47311a9c-74d2-400d-a22d-9ef55b14edb5 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1040.703074] env[62385]: DEBUG nova.network.neutron [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Updating instance_info_cache with network_info: [{"id": "b19d5597-14de-4b08-b628-3bc8e539f77a", "address": "fa:16:3e:83:b5:c7", "network": {"id": "a561a523-c9d8-4169-b449-5c0bdef7b53b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-751953866-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d716d19f79945db98f684df0cf4b302", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb19d5597-14", "ovs_interfaceid": "b19d5597-14de-4b08-b628-3bc8e539f77a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.778345] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "5c9284ba-9f83-454f-b91b-9b0ea9ee3195" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.778580] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "5c9284ba-9f83-454f-b91b-9b0ea9ee3195" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.908894] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206394, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.114352] env[62385]: DEBUG nova.objects.instance [None req-56e99cef-0ff6-4dac-b865-2525195eae09 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lazy-loading 'flavor' on Instance uuid a1c71d8b-0953-4cc2-961d-0c2b2fc11110 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1041.148617] env[62385]: DEBUG nova.compute.utils [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1041.150069] env[62385]: DEBUG nova.compute.manager [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1041.150198] env[62385]: DEBUG nova.network.neutron [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1041.192588] env[62385]: DEBUG nova.policy [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7e04c6591fe043d686140eb2f5ce0be2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd45b5346eab44707b4d81eaaf12b0935', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 1041.205908] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Releasing lock "refresh_cache-c49d4366-2ea0-44f2-8463-daf62bd1c40c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.206247] env[62385]: DEBUG nova.compute.manager [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Instance network_info: |[{"id": "b19d5597-14de-4b08-b628-3bc8e539f77a", "address": "fa:16:3e:83:b5:c7", "network": {"id": "a561a523-c9d8-4169-b449-5c0bdef7b53b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-751953866-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d716d19f79945db98f684df0cf4b302", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb19d5597-14", "ovs_interfaceid": "b19d5597-14de-4b08-b628-3bc8e539f77a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1041.208954] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:83:b5:c7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '170f3b82-5915-4e36-bce9-4664ebb6be5e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b19d5597-14de-4b08-b628-3bc8e539f77a', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1041.217501] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Creating folder: Project (5d716d19f79945db98f684df0cf4b302). Parent ref: group-v261107. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1041.218058] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ada6841d-1f1f-46ab-88c6-2a531eaf8baf {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.233749] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Created folder: Project (5d716d19f79945db98f684df0cf4b302) in parent group-v261107. [ 1041.234028] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Creating folder: Instances. Parent ref: group-v261301. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1041.234416] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d28cbf81-8260-4a7f-aff2-a4d126940cdf {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.249586] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Created folder: Instances in parent group-v261301. [ 1041.249858] env[62385]: DEBUG oslo.service.loopingcall [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1041.250074] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1041.250310] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-98bceefe-e262-4802-90fa-1a6635213891 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.272788] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1041.272788] env[62385]: value = "task-1206399" [ 1041.272788] env[62385]: _type = "Task" [ 1041.272788] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.282740] env[62385]: DEBUG nova.compute.manager [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1041.285185] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206399, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.313787] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68ebf1d1-3ab0-47a1-908b-c8eaf0f08862 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.322180] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa56fd5b-748a-4338-8603-8ea6bc7bdb18 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.352963] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef0d7d40-01db-4b98-84ab-452b5bd04d9d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.361544] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37ce59d7-a8be-436b-9e69-39e7dee10fad {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.382155] env[62385]: DEBUG nova.compute.provider_tree [None req-821932eb-b83d-456c-bbce-fef50c0c3b8b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1041.411436] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206394, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.357729} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.411813] env[62385]: INFO nova.virt.vmwareapi.ds_util [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_dc01d2d7-f982-46ac-bca5-1b922f01ffe5/OSTACK_IMG_dc01d2d7-f982-46ac-bca5-1b922f01ffe5.vmdk to [datastore1] devstack-image-cache_base/a561779b-d9f7-4373-8a4d-586fbc25c6d7/a561779b-d9f7-4373-8a4d-586fbc25c6d7.vmdk. [ 1041.412104] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Cleaning up location [datastore1] OSTACK_IMG_dc01d2d7-f982-46ac-bca5-1b922f01ffe5 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1041.412359] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_dc01d2d7-f982-46ac-bca5-1b922f01ffe5 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1041.412702] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-31e459d8-848d-43dd-b5bd-70d550174e41 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.420487] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1041.420487] env[62385]: value = "task-1206400" [ 1041.420487] env[62385]: _type = "Task" [ 1041.420487] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.428578] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206400, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.494630] env[62385]: DEBUG nova.compute.manager [req-d4606ed6-9941-4e6d-88e2-0f6a8ef5681e req-6d4047f4-973c-4118-adee-96640fc03376 service nova] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Received event network-changed-b19d5597-14de-4b08-b628-3bc8e539f77a {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1041.494842] env[62385]: DEBUG nova.compute.manager [req-d4606ed6-9941-4e6d-88e2-0f6a8ef5681e req-6d4047f4-973c-4118-adee-96640fc03376 service nova] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Refreshing instance network info cache due to event network-changed-b19d5597-14de-4b08-b628-3bc8e539f77a. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1041.495078] env[62385]: DEBUG oslo_concurrency.lockutils [req-d4606ed6-9941-4e6d-88e2-0f6a8ef5681e req-6d4047f4-973c-4118-adee-96640fc03376 service nova] Acquiring lock "refresh_cache-c49d4366-2ea0-44f2-8463-daf62bd1c40c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1041.495229] env[62385]: DEBUG oslo_concurrency.lockutils [req-d4606ed6-9941-4e6d-88e2-0f6a8ef5681e req-6d4047f4-973c-4118-adee-96640fc03376 service nova] Acquired lock "refresh_cache-c49d4366-2ea0-44f2-8463-daf62bd1c40c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.495391] env[62385]: DEBUG nova.network.neutron [req-d4606ed6-9941-4e6d-88e2-0f6a8ef5681e req-6d4047f4-973c-4118-adee-96640fc03376 service nova] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Refreshing network info cache for port b19d5597-14de-4b08-b628-3bc8e539f77a {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1041.565564] env[62385]: DEBUG nova.network.neutron [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Successfully created port: c6ede9da-b209-4fae-b75a-c2f1ede4dcaf {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1041.656297] env[62385]: DEBUG nova.compute.manager [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1041.783877] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206399, 'name': CreateVM_Task} progress is 25%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.800871] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.885964] env[62385]: DEBUG nova.scheduler.client.report [None req-821932eb-b83d-456c-bbce-fef50c0c3b8b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1041.930419] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206400, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.185116} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.930713] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1041.930869] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a561779b-d9f7-4373-8a4d-586fbc25c6d7/a561779b-d9f7-4373-8a4d-586fbc25c6d7.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.931164] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a561779b-d9f7-4373-8a4d-586fbc25c6d7/a561779b-d9f7-4373-8a4d-586fbc25c6d7.vmdk to [datastore1] 8e1d416c-2624-468c-94d7-6265e9f4178f/8e1d416c-2624-468c-94d7-6265e9f4178f.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1041.931460] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8b04dffb-7e9f-4078-ae95-e3074996c8f3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.939566] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1041.939566] env[62385]: value = "task-1206401" [ 1041.939566] env[62385]: _type = "Task" [ 1041.939566] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.948813] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206401, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.123247] env[62385]: DEBUG oslo_concurrency.lockutils [None req-56e99cef-0ff6-4dac-b865-2525195eae09 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "a1c71d8b-0953-4cc2-961d-0c2b2fc11110" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.764s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.195924] env[62385]: DEBUG nova.network.neutron [None req-fde088e5-3eb5-4961-b95b-023de5dbe558 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Successfully updated port: 331b6188-b103-437e-8436-5259bc356e31 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1042.211900] env[62385]: DEBUG nova.network.neutron [req-d4606ed6-9941-4e6d-88e2-0f6a8ef5681e req-6d4047f4-973c-4118-adee-96640fc03376 service nova] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Updated VIF entry in instance network info cache for port b19d5597-14de-4b08-b628-3bc8e539f77a. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1042.212470] env[62385]: DEBUG nova.network.neutron [req-d4606ed6-9941-4e6d-88e2-0f6a8ef5681e req-6d4047f4-973c-4118-adee-96640fc03376 service nova] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Updating instance_info_cache with network_info: [{"id": "b19d5597-14de-4b08-b628-3bc8e539f77a", "address": "fa:16:3e:83:b5:c7", "network": {"id": "a561a523-c9d8-4169-b449-5c0bdef7b53b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-751953866-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d716d19f79945db98f684df0cf4b302", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb19d5597-14", "ovs_interfaceid": "b19d5597-14de-4b08-b628-3bc8e539f77a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.284956] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206399, 'name': CreateVM_Task, 'duration_secs': 0.839599} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.285156] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1042.285892] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.286084] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.286434] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1042.286711] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6aecc167-ed2f-4f9a-9a0f-d71a805f818e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.292732] env[62385]: DEBUG oslo_vmware.api [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for the task: (returnval){ [ 1042.292732] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52bb4070-78c9-5b7e-2b91-12e5a9d35b65" [ 1042.292732] env[62385]: _type = "Task" [ 1042.292732] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.303628] env[62385]: DEBUG oslo_vmware.api [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52bb4070-78c9-5b7e-2b91-12e5a9d35b65, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.392736] env[62385]: DEBUG oslo_concurrency.lockutils [None req-821932eb-b83d-456c-bbce-fef50c0c3b8b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.747s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.395799] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.595s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.397738] env[62385]: INFO nova.compute.claims [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1042.417775] env[62385]: INFO nova.scheduler.client.report [None req-821932eb-b83d-456c-bbce-fef50c0c3b8b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Deleted allocations for instance 47311a9c-74d2-400d-a22d-9ef55b14edb5 [ 1042.451190] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206401, 'name': CopyVirtualDisk_Task} progress is 15%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.658769] env[62385]: DEBUG oslo_concurrency.lockutils [None req-14e8ec1a-422a-4b0e-aa3f-2d460d7de55c tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Acquiring lock "a1c71d8b-0953-4cc2-961d-0c2b2fc11110" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.659012] env[62385]: DEBUG oslo_concurrency.lockutils [None req-14e8ec1a-422a-4b0e-aa3f-2d460d7de55c tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "a1c71d8b-0953-4cc2-961d-0c2b2fc11110" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.667438] env[62385]: DEBUG nova.compute.manager [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1042.697415] env[62385]: DEBUG nova.virt.hardware [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1042.697670] env[62385]: DEBUG nova.virt.hardware [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1042.697835] env[62385]: DEBUG nova.virt.hardware [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1042.698212] env[62385]: DEBUG nova.virt.hardware [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1042.698503] env[62385]: DEBUG nova.virt.hardware [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1042.698695] env[62385]: DEBUG nova.virt.hardware [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1042.699362] env[62385]: DEBUG nova.virt.hardware [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1042.699642] env[62385]: DEBUG nova.virt.hardware [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1042.699977] env[62385]: DEBUG nova.virt.hardware [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1042.700401] env[62385]: DEBUG nova.virt.hardware [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1042.700693] env[62385]: DEBUG nova.virt.hardware [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1042.701660] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fde088e5-3eb5-4961-b95b-023de5dbe558 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "refresh_cache-57637d6e-39e2-405c-bbcf-64bf0ffbebd1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.701904] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fde088e5-3eb5-4961-b95b-023de5dbe558 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquired lock "refresh_cache-57637d6e-39e2-405c-bbcf-64bf0ffbebd1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.702164] env[62385]: DEBUG nova.network.neutron [None req-fde088e5-3eb5-4961-b95b-023de5dbe558 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1042.704108] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6a9b882-09d3-4868-b1b4-166a43141090 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.715134] env[62385]: DEBUG oslo_concurrency.lockutils [req-d4606ed6-9941-4e6d-88e2-0f6a8ef5681e req-6d4047f4-973c-4118-adee-96640fc03376 service nova] Releasing lock "refresh_cache-c49d4366-2ea0-44f2-8463-daf62bd1c40c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1042.720195] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6e640eb-1ad5-4270-b1e2-35cc5b101cd4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.803685] env[62385]: DEBUG oslo_vmware.api [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52bb4070-78c9-5b7e-2b91-12e5a9d35b65, 'name': SearchDatastore_Task, 'duration_secs': 0.051018} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.804050] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1042.804304] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1042.804552] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.804736] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.804882] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1042.805177] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-831448e9-5ed9-4a1e-a117-339cf3fe6a16 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.821841] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1042.822138] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1042.823032] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ccad3351-91b6-43aa-9a43-8ec2c0b3a0c2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.829275] env[62385]: DEBUG oslo_vmware.api [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for the task: (returnval){ [ 1042.829275] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52760612-1f09-87f3-3511-2ca5a5d78f7e" [ 1042.829275] env[62385]: _type = "Task" [ 1042.829275] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.838178] env[62385]: DEBUG oslo_vmware.api [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52760612-1f09-87f3-3511-2ca5a5d78f7e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.929136] env[62385]: DEBUG oslo_concurrency.lockutils [None req-821932eb-b83d-456c-bbce-fef50c0c3b8b tempest-ServersTestJSON-792525361 tempest-ServersTestJSON-792525361-project-member] Lock "47311a9c-74d2-400d-a22d-9ef55b14edb5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.358s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.951276] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206401, 'name': CopyVirtualDisk_Task} progress is 38%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.162717] env[62385]: INFO nova.compute.manager [None req-14e8ec1a-422a-4b0e-aa3f-2d460d7de55c tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Detaching volume 67e47839-a66b-4032-b1c4-93171c686c2f [ 1043.207923] env[62385]: INFO nova.virt.block_device [None req-14e8ec1a-422a-4b0e-aa3f-2d460d7de55c tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Attempting to driver detach volume 67e47839-a66b-4032-b1c4-93171c686c2f from mountpoint /dev/sdc [ 1043.208320] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-14e8ec1a-422a-4b0e-aa3f-2d460d7de55c tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Volume detach. Driver type: vmdk {{(pid=62385) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1043.208641] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-14e8ec1a-422a-4b0e-aa3f-2d460d7de55c tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261298', 'volume_id': '67e47839-a66b-4032-b1c4-93171c686c2f', 'name': 'volume-67e47839-a66b-4032-b1c4-93171c686c2f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a1c71d8b-0953-4cc2-961d-0c2b2fc11110', 'attached_at': '', 'detached_at': '', 'volume_id': '67e47839-a66b-4032-b1c4-93171c686c2f', 'serial': '67e47839-a66b-4032-b1c4-93171c686c2f'} {{(pid=62385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1043.214638] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ef27977-a4cc-4a47-9730-50c4c2e61550 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.244656] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67e51b9f-5750-42f1-a949-6edfaf23fde2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.256255] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e27248fe-29ec-4767-a2ab-296646bd1d6f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.279207] env[62385]: WARNING nova.network.neutron [None req-fde088e5-3eb5-4961-b95b-023de5dbe558 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] daad7070-dc06-4580-9f4a-ac8aaa546955 already exists in list: networks containing: ['daad7070-dc06-4580-9f4a-ac8aaa546955']. ignoring it [ 1043.281787] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87aca414-b6bb-4992-bfec-b21dc9d06bee {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.302300] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-14e8ec1a-422a-4b0e-aa3f-2d460d7de55c tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] The volume has not been displaced from its original location: [datastore1] volume-67e47839-a66b-4032-b1c4-93171c686c2f/volume-67e47839-a66b-4032-b1c4-93171c686c2f.vmdk. No consolidation needed. {{(pid=62385) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1043.307500] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-14e8ec1a-422a-4b0e-aa3f-2d460d7de55c tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Reconfiguring VM instance instance-00000063 to detach disk 2002 {{(pid=62385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1043.308045] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1582b4fd-ab26-49bb-b248-8e7c51eb533f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.328077] env[62385]: DEBUG oslo_vmware.api [None req-14e8ec1a-422a-4b0e-aa3f-2d460d7de55c tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for the task: (returnval){ [ 1043.328077] env[62385]: value = "task-1206402" [ 1043.328077] env[62385]: _type = "Task" [ 1043.328077] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.340695] env[62385]: DEBUG oslo_vmware.api [None req-14e8ec1a-422a-4b0e-aa3f-2d460d7de55c tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206402, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.344755] env[62385]: DEBUG oslo_vmware.api [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52760612-1f09-87f3-3511-2ca5a5d78f7e, 'name': SearchDatastore_Task, 'duration_secs': 0.084605} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.345872] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e41e269-3058-4f0e-a338-da82b3d8fa5a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.354339] env[62385]: DEBUG oslo_vmware.api [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for the task: (returnval){ [ 1043.354339] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5239ed9e-b312-2fa4-6f42-576e9f092d28" [ 1043.354339] env[62385]: _type = "Task" [ 1043.354339] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.369438] env[62385]: DEBUG oslo_vmware.api [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5239ed9e-b312-2fa4-6f42-576e9f092d28, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.460019] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206401, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.532371] env[62385]: DEBUG nova.compute.manager [req-27222813-2200-4279-855c-6a291b649986 req-5e172c09-8871-43d6-a4a6-10fa3e9abaa8 service nova] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Received event network-vif-plugged-331b6188-b103-437e-8436-5259bc356e31 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1043.532371] env[62385]: DEBUG oslo_concurrency.lockutils [req-27222813-2200-4279-855c-6a291b649986 req-5e172c09-8871-43d6-a4a6-10fa3e9abaa8 service nova] Acquiring lock "57637d6e-39e2-405c-bbcf-64bf0ffbebd1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.532717] env[62385]: DEBUG oslo_concurrency.lockutils [req-27222813-2200-4279-855c-6a291b649986 req-5e172c09-8871-43d6-a4a6-10fa3e9abaa8 service nova] Lock "57637d6e-39e2-405c-bbcf-64bf0ffbebd1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.533156] env[62385]: DEBUG oslo_concurrency.lockutils [req-27222813-2200-4279-855c-6a291b649986 req-5e172c09-8871-43d6-a4a6-10fa3e9abaa8 service nova] Lock "57637d6e-39e2-405c-bbcf-64bf0ffbebd1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.533383] env[62385]: DEBUG nova.compute.manager [req-27222813-2200-4279-855c-6a291b649986 req-5e172c09-8871-43d6-a4a6-10fa3e9abaa8 service nova] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] No waiting events found dispatching network-vif-plugged-331b6188-b103-437e-8436-5259bc356e31 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1043.533610] env[62385]: WARNING nova.compute.manager [req-27222813-2200-4279-855c-6a291b649986 req-5e172c09-8871-43d6-a4a6-10fa3e9abaa8 service nova] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Received unexpected event network-vif-plugged-331b6188-b103-437e-8436-5259bc356e31 for instance with vm_state active and task_state None. [ 1043.533806] env[62385]: DEBUG nova.compute.manager [req-27222813-2200-4279-855c-6a291b649986 req-5e172c09-8871-43d6-a4a6-10fa3e9abaa8 service nova] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Received event network-changed-331b6188-b103-437e-8436-5259bc356e31 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1043.533970] env[62385]: DEBUG nova.compute.manager [req-27222813-2200-4279-855c-6a291b649986 req-5e172c09-8871-43d6-a4a6-10fa3e9abaa8 service nova] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Refreshing instance network info cache due to event network-changed-331b6188-b103-437e-8436-5259bc356e31. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1043.534558] env[62385]: DEBUG oslo_concurrency.lockutils [req-27222813-2200-4279-855c-6a291b649986 req-5e172c09-8871-43d6-a4a6-10fa3e9abaa8 service nova] Acquiring lock "refresh_cache-57637d6e-39e2-405c-bbcf-64bf0ffbebd1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.570270] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-106eba5d-3e27-4290-9ebe-0fe9433473e2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.581420] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f219e9c-0a20-4678-abdc-b85ae253d26f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.623865] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d201748a-c31a-49b8-b9f6-cb4beae0c6d7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.635085] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0c96c21-a2cf-4281-aa16-e2d16a674946 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.654335] env[62385]: DEBUG nova.compute.provider_tree [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1043.666403] env[62385]: DEBUG nova.network.neutron [None req-fde088e5-3eb5-4961-b95b-023de5dbe558 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Updating instance_info_cache with network_info: [{"id": "a7153a2d-db69-4d14-8ddc-b634ddabbc18", "address": "fa:16:3e:a3:0d:ab", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.210", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7153a2d-db", "ovs_interfaceid": "a7153a2d-db69-4d14-8ddc-b634ddabbc18", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "331b6188-b103-437e-8436-5259bc356e31", "address": "fa:16:3e:f4:84:82", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap331b6188-b1", "ovs_interfaceid": "331b6188-b103-437e-8436-5259bc356e31", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.841343] env[62385]: DEBUG oslo_vmware.api [None req-14e8ec1a-422a-4b0e-aa3f-2d460d7de55c tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206402, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.869710] env[62385]: DEBUG oslo_vmware.api [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5239ed9e-b312-2fa4-6f42-576e9f092d28, 'name': SearchDatastore_Task, 'duration_secs': 0.097788} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.869995] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.870277] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] c49d4366-2ea0-44f2-8463-daf62bd1c40c/c49d4366-2ea0-44f2-8463-daf62bd1c40c.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1043.870600] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-67f029f3-0479-4634-bfa0-4cdf29ad68c8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.881279] env[62385]: DEBUG oslo_vmware.api [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for the task: (returnval){ [ 1043.881279] env[62385]: value = "task-1206403" [ 1043.881279] env[62385]: _type = "Task" [ 1043.881279] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.892826] env[62385]: DEBUG oslo_vmware.api [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206403, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.957497] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206401, 'name': CopyVirtualDisk_Task} progress is 80%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.148905] env[62385]: DEBUG nova.network.neutron [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Successfully updated port: c6ede9da-b209-4fae-b75a-c2f1ede4dcaf {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1044.157740] env[62385]: DEBUG nova.scheduler.client.report [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1044.168648] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fde088e5-3eb5-4961-b95b-023de5dbe558 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Releasing lock "refresh_cache-57637d6e-39e2-405c-bbcf-64bf0ffbebd1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.169640] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fde088e5-3eb5-4961-b95b-023de5dbe558 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "57637d6e-39e2-405c-bbcf-64bf0ffbebd1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.169725] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fde088e5-3eb5-4961-b95b-023de5dbe558 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquired lock "57637d6e-39e2-405c-bbcf-64bf0ffbebd1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.171250] env[62385]: DEBUG oslo_concurrency.lockutils [req-27222813-2200-4279-855c-6a291b649986 req-5e172c09-8871-43d6-a4a6-10fa3e9abaa8 service nova] Acquired lock "refresh_cache-57637d6e-39e2-405c-bbcf-64bf0ffbebd1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.171250] env[62385]: DEBUG nova.network.neutron [req-27222813-2200-4279-855c-6a291b649986 req-5e172c09-8871-43d6-a4a6-10fa3e9abaa8 service nova] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Refreshing network info cache for port 331b6188-b103-437e-8436-5259bc356e31 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1044.178715] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9741e142-bbd1-48af-84ef-64fe21c848e6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.199795] env[62385]: DEBUG nova.virt.hardware [None req-fde088e5-3eb5-4961-b95b-023de5dbe558 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1044.200132] env[62385]: DEBUG nova.virt.hardware [None req-fde088e5-3eb5-4961-b95b-023de5dbe558 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1044.200387] env[62385]: DEBUG nova.virt.hardware [None req-fde088e5-3eb5-4961-b95b-023de5dbe558 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1044.200612] env[62385]: DEBUG nova.virt.hardware [None req-fde088e5-3eb5-4961-b95b-023de5dbe558 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1044.200836] env[62385]: DEBUG nova.virt.hardware [None req-fde088e5-3eb5-4961-b95b-023de5dbe558 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1044.201068] env[62385]: DEBUG nova.virt.hardware [None req-fde088e5-3eb5-4961-b95b-023de5dbe558 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1044.201773] env[62385]: DEBUG nova.virt.hardware [None req-fde088e5-3eb5-4961-b95b-023de5dbe558 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1044.201773] env[62385]: DEBUG nova.virt.hardware [None req-fde088e5-3eb5-4961-b95b-023de5dbe558 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1044.201843] env[62385]: DEBUG nova.virt.hardware [None req-fde088e5-3eb5-4961-b95b-023de5dbe558 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1044.201996] env[62385]: DEBUG nova.virt.hardware [None req-fde088e5-3eb5-4961-b95b-023de5dbe558 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1044.202230] env[62385]: DEBUG nova.virt.hardware [None req-fde088e5-3eb5-4961-b95b-023de5dbe558 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1044.210714] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fde088e5-3eb5-4961-b95b-023de5dbe558 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Reconfiguring VM to attach interface {{(pid=62385) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 1044.211582] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-497b844b-afef-40a7-9878-d76d9869ec9e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.234517] env[62385]: DEBUG oslo_vmware.api [None req-fde088e5-3eb5-4961-b95b-023de5dbe558 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 1044.234517] env[62385]: value = "task-1206404" [ 1044.234517] env[62385]: _type = "Task" [ 1044.234517] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.251091] env[62385]: DEBUG oslo_vmware.api [None req-fde088e5-3eb5-4961-b95b-023de5dbe558 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206404, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.344469] env[62385]: DEBUG oslo_vmware.api [None req-14e8ec1a-422a-4b0e-aa3f-2d460d7de55c tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206402, 'name': ReconfigVM_Task, 'duration_secs': 0.627818} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.344675] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-14e8ec1a-422a-4b0e-aa3f-2d460d7de55c tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Reconfigured VM instance instance-00000063 to detach disk 2002 {{(pid=62385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1044.350141] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-be2d4e8b-e1a8-4a2e-87ff-0a984424949c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.371359] env[62385]: DEBUG oslo_vmware.api [None req-14e8ec1a-422a-4b0e-aa3f-2d460d7de55c tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for the task: (returnval){ [ 1044.371359] env[62385]: value = "task-1206405" [ 1044.371359] env[62385]: _type = "Task" [ 1044.371359] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.383336] env[62385]: DEBUG oslo_vmware.api [None req-14e8ec1a-422a-4b0e-aa3f-2d460d7de55c tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206405, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.394290] env[62385]: DEBUG oslo_vmware.api [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206403, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.454865] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206401, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.490537} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.455306] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a561779b-d9f7-4373-8a4d-586fbc25c6d7/a561779b-d9f7-4373-8a4d-586fbc25c6d7.vmdk to [datastore1] 8e1d416c-2624-468c-94d7-6265e9f4178f/8e1d416c-2624-468c-94d7-6265e9f4178f.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1044.456195] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e31bb7d6-dc01-4a27-9757-ea4247b41a86 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.484428] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] 8e1d416c-2624-468c-94d7-6265e9f4178f/8e1d416c-2624-468c-94d7-6265e9f4178f.vmdk or device None with type streamOptimized {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1044.484985] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b5e89c90-b46b-41f6-b521-0905aaae11fe {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.509862] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1044.509862] env[62385]: value = "task-1206406" [ 1044.509862] env[62385]: _type = "Task" [ 1044.509862] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.519622] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206406, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.651927] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "refresh_cache-1289e63a-5a9d-4615-b6cf-bbf5f4ae0317" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.652122] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquired lock "refresh_cache-1289e63a-5a9d-4615-b6cf-bbf5f4ae0317" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.652300] env[62385]: DEBUG nova.network.neutron [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1044.668264] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.270s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.668264] env[62385]: DEBUG nova.compute.manager [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1044.747734] env[62385]: DEBUG oslo_vmware.api [None req-fde088e5-3eb5-4961-b95b-023de5dbe558 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206404, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.886189] env[62385]: DEBUG oslo_vmware.api [None req-14e8ec1a-422a-4b0e-aa3f-2d460d7de55c tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206405, 'name': ReconfigVM_Task, 'duration_secs': 0.18544} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.890484] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-14e8ec1a-422a-4b0e-aa3f-2d460d7de55c tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261298', 'volume_id': '67e47839-a66b-4032-b1c4-93171c686c2f', 'name': 'volume-67e47839-a66b-4032-b1c4-93171c686c2f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a1c71d8b-0953-4cc2-961d-0c2b2fc11110', 'attached_at': '', 'detached_at': '', 'volume_id': '67e47839-a66b-4032-b1c4-93171c686c2f', 'serial': '67e47839-a66b-4032-b1c4-93171c686c2f'} {{(pid=62385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1044.901240] env[62385]: DEBUG oslo_vmware.api [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206403, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.022893] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206406, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.124817] env[62385]: DEBUG nova.network.neutron [req-27222813-2200-4279-855c-6a291b649986 req-5e172c09-8871-43d6-a4a6-10fa3e9abaa8 service nova] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Updated VIF entry in instance network info cache for port 331b6188-b103-437e-8436-5259bc356e31. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1045.127137] env[62385]: DEBUG nova.network.neutron [req-27222813-2200-4279-855c-6a291b649986 req-5e172c09-8871-43d6-a4a6-10fa3e9abaa8 service nova] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Updating instance_info_cache with network_info: [{"id": "a7153a2d-db69-4d14-8ddc-b634ddabbc18", "address": "fa:16:3e:a3:0d:ab", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.210", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7153a2d-db", "ovs_interfaceid": "a7153a2d-db69-4d14-8ddc-b634ddabbc18", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "331b6188-b103-437e-8436-5259bc356e31", "address": "fa:16:3e:f4:84:82", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap331b6188-b1", "ovs_interfaceid": "331b6188-b103-437e-8436-5259bc356e31", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.174620] env[62385]: DEBUG nova.compute.utils [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1045.176115] env[62385]: DEBUG nova.compute.manager [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1045.176314] env[62385]: DEBUG nova.network.neutron [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1045.214949] env[62385]: DEBUG nova.network.neutron [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1045.232767] env[62385]: DEBUG nova.policy [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b314c392b9ca405480c27c25364fda81', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '93111f5ba17c47de9691f90b99e03aa8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 1045.249221] env[62385]: DEBUG oslo_vmware.api [None req-fde088e5-3eb5-4961-b95b-023de5dbe558 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206404, 'name': ReconfigVM_Task, 'duration_secs': 0.96327} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.249986] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fde088e5-3eb5-4961-b95b-023de5dbe558 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Releasing lock "57637d6e-39e2-405c-bbcf-64bf0ffbebd1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.249986] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-fde088e5-3eb5-4961-b95b-023de5dbe558 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Reconfigured VM to attach interface {{(pid=62385) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 1045.400618] env[62385]: DEBUG oslo_vmware.api [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206403, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.058359} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.401150] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] c49d4366-2ea0-44f2-8463-daf62bd1c40c/c49d4366-2ea0-44f2-8463-daf62bd1c40c.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1045.401427] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1045.401691] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1da955d3-bfd0-47ed-9a23-b91e4dbcbfb1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.410636] env[62385]: DEBUG oslo_vmware.api [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for the task: (returnval){ [ 1045.410636] env[62385]: value = "task-1206407" [ 1045.410636] env[62385]: _type = "Task" [ 1045.410636] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.424110] env[62385]: DEBUG oslo_vmware.api [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206407, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.445414] env[62385]: DEBUG nova.objects.instance [None req-14e8ec1a-422a-4b0e-aa3f-2d460d7de55c tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lazy-loading 'flavor' on Instance uuid a1c71d8b-0953-4cc2-961d-0c2b2fc11110 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1045.466802] env[62385]: DEBUG nova.network.neutron [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Updating instance_info_cache with network_info: [{"id": "c6ede9da-b209-4fae-b75a-c2f1ede4dcaf", "address": "fa:16:3e:a7:e1:c0", "network": {"id": "06017161-984c-4657-9639-4e6bdb6d4c09", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1072882119-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d45b5346eab44707b4d81eaaf12b0935", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39cd75b0-9ec7-48ed-b57f-34da0c573a60", "external-id": "nsx-vlan-transportzone-751", "segmentation_id": 751, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6ede9da-b2", "ovs_interfaceid": "c6ede9da-b209-4fae-b75a-c2f1ede4dcaf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.520840] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206406, 'name': ReconfigVM_Task, 'duration_secs': 0.764758} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.521191] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Reconfigured VM instance instance-0000005b to attach disk [datastore1] 8e1d416c-2624-468c-94d7-6265e9f4178f/8e1d416c-2624-468c-94d7-6265e9f4178f.vmdk or device None with type streamOptimized {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1045.522292] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'size': 0, 'encrypted': False, 'disk_bus': None, 'encryption_secret_uuid': None, 'boot_index': 0, 'encryption_options': None, 'device_type': 'disk', 'guest_format': None, 'device_name': '/dev/sda', 'encryption_format': None, 'image_id': 'c0abbb8d-77e7-4f22-a256-0faf0a781109'}], 'ephemerals': [], 'block_device_mapping': [{'disk_bus': None, 'attachment_id': 'd3b7c0c2-0351-4826-ad5c-f584c6d8097c', 'boot_index': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261292', 'volume_id': '925fe844-0d8f-4668-ac82-328339e88cac', 'name': 'volume-925fe844-0d8f-4668-ac82-328339e88cac', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '8e1d416c-2624-468c-94d7-6265e9f4178f', 'attached_at': '', 'detached_at': '', 'volume_id': '925fe844-0d8f-4668-ac82-328339e88cac', 'serial': '925fe844-0d8f-4668-ac82-328339e88cac'}, 'delete_on_termination': False, 'mount_device': '/dev/sdb', 'device_type': None, 'guest_format': None, 'volume_type': None}], 'swap': None} {{(pid=62385) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1045.522503] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Volume attach. Driver type: vmdk {{(pid=62385) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1045.523144] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261292', 'volume_id': '925fe844-0d8f-4668-ac82-328339e88cac', 'name': 'volume-925fe844-0d8f-4668-ac82-328339e88cac', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '8e1d416c-2624-468c-94d7-6265e9f4178f', 'attached_at': '', 'detached_at': '', 'volume_id': '925fe844-0d8f-4668-ac82-328339e88cac', 'serial': '925fe844-0d8f-4668-ac82-328339e88cac'} {{(pid=62385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1045.523495] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18f21bfe-4ade-460c-bda3-f44105cf2881 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.540746] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d3401e1-3146-4a4b-89bd-569e8daa4402 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.571669] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] volume-925fe844-0d8f-4668-ac82-328339e88cac/volume-925fe844-0d8f-4668-ac82-328339e88cac.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1045.572934] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-075e446d-95cd-40b1-9fcd-1d0af5cd31ac {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.586423] env[62385]: DEBUG nova.compute.manager [req-0c8b2307-de0e-4e29-930b-e06f1a97a83c req-92de5191-da96-4de3-ab10-6ef083724438 service nova] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Received event network-vif-plugged-c6ede9da-b209-4fae-b75a-c2f1ede4dcaf {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1045.586641] env[62385]: DEBUG oslo_concurrency.lockutils [req-0c8b2307-de0e-4e29-930b-e06f1a97a83c req-92de5191-da96-4de3-ab10-6ef083724438 service nova] Acquiring lock "1289e63a-5a9d-4615-b6cf-bbf5f4ae0317-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.587202] env[62385]: DEBUG oslo_concurrency.lockutils [req-0c8b2307-de0e-4e29-930b-e06f1a97a83c req-92de5191-da96-4de3-ab10-6ef083724438 service nova] Lock "1289e63a-5a9d-4615-b6cf-bbf5f4ae0317-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.587760] env[62385]: DEBUG oslo_concurrency.lockutils [req-0c8b2307-de0e-4e29-930b-e06f1a97a83c req-92de5191-da96-4de3-ab10-6ef083724438 service nova] Lock "1289e63a-5a9d-4615-b6cf-bbf5f4ae0317-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.587760] env[62385]: DEBUG nova.compute.manager [req-0c8b2307-de0e-4e29-930b-e06f1a97a83c req-92de5191-da96-4de3-ab10-6ef083724438 service nova] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] No waiting events found dispatching network-vif-plugged-c6ede9da-b209-4fae-b75a-c2f1ede4dcaf {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1045.587990] env[62385]: WARNING nova.compute.manager [req-0c8b2307-de0e-4e29-930b-e06f1a97a83c req-92de5191-da96-4de3-ab10-6ef083724438 service nova] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Received unexpected event network-vif-plugged-c6ede9da-b209-4fae-b75a-c2f1ede4dcaf for instance with vm_state building and task_state spawning. [ 1045.588051] env[62385]: DEBUG nova.compute.manager [req-0c8b2307-de0e-4e29-930b-e06f1a97a83c req-92de5191-da96-4de3-ab10-6ef083724438 service nova] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Received event network-changed-c6ede9da-b209-4fae-b75a-c2f1ede4dcaf {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1045.588239] env[62385]: DEBUG nova.compute.manager [req-0c8b2307-de0e-4e29-930b-e06f1a97a83c req-92de5191-da96-4de3-ab10-6ef083724438 service nova] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Refreshing instance network info cache due to event network-changed-c6ede9da-b209-4fae-b75a-c2f1ede4dcaf. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1045.588419] env[62385]: DEBUG oslo_concurrency.lockutils [req-0c8b2307-de0e-4e29-930b-e06f1a97a83c req-92de5191-da96-4de3-ab10-6ef083724438 service nova] Acquiring lock "refresh_cache-1289e63a-5a9d-4615-b6cf-bbf5f4ae0317" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.594912] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1045.594912] env[62385]: value = "task-1206408" [ 1045.594912] env[62385]: _type = "Task" [ 1045.594912] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.603283] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206408, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.636238] env[62385]: DEBUG oslo_concurrency.lockutils [req-27222813-2200-4279-855c-6a291b649986 req-5e172c09-8871-43d6-a4a6-10fa3e9abaa8 service nova] Releasing lock "refresh_cache-57637d6e-39e2-405c-bbcf-64bf0ffbebd1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.686674] env[62385]: DEBUG nova.compute.manager [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1045.756281] env[62385]: DEBUG oslo_concurrency.lockutils [None req-fde088e5-3eb5-4961-b95b-023de5dbe558 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "interface-57637d6e-39e2-405c-bbcf-64bf0ffbebd1-331b6188-b103-437e-8436-5259bc356e31" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.733s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.772663] env[62385]: DEBUG nova.network.neutron [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Successfully created port: b5615be2-bc51-4fe3-8905-4b7083550949 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1045.928435] env[62385]: DEBUG oslo_vmware.api [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206407, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.10338} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.928435] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1045.928435] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06c53334-bd81-42fe-b070-8c574c3eb5ef {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.954012] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] c49d4366-2ea0-44f2-8463-daf62bd1c40c/c49d4366-2ea0-44f2-8463-daf62bd1c40c.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1045.958424] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-60b436b1-29d3-4bc8-825e-d0bcdfe670cb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.977144] env[62385]: DEBUG oslo_concurrency.lockutils [None req-14e8ec1a-422a-4b0e-aa3f-2d460d7de55c tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "a1c71d8b-0953-4cc2-961d-0c2b2fc11110" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.318s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.978577] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Releasing lock "refresh_cache-1289e63a-5a9d-4615-b6cf-bbf5f4ae0317" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.979085] env[62385]: DEBUG nova.compute.manager [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Instance network_info: |[{"id": "c6ede9da-b209-4fae-b75a-c2f1ede4dcaf", "address": "fa:16:3e:a7:e1:c0", "network": {"id": "06017161-984c-4657-9639-4e6bdb6d4c09", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1072882119-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d45b5346eab44707b4d81eaaf12b0935", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39cd75b0-9ec7-48ed-b57f-34da0c573a60", "external-id": "nsx-vlan-transportzone-751", "segmentation_id": 751, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6ede9da-b2", "ovs_interfaceid": "c6ede9da-b209-4fae-b75a-c2f1ede4dcaf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1045.979643] env[62385]: DEBUG oslo_concurrency.lockutils [req-0c8b2307-de0e-4e29-930b-e06f1a97a83c req-92de5191-da96-4de3-ab10-6ef083724438 service nova] Acquired lock "refresh_cache-1289e63a-5a9d-4615-b6cf-bbf5f4ae0317" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.979831] env[62385]: DEBUG nova.network.neutron [req-0c8b2307-de0e-4e29-930b-e06f1a97a83c req-92de5191-da96-4de3-ab10-6ef083724438 service nova] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Refreshing network info cache for port c6ede9da-b209-4fae-b75a-c2f1ede4dcaf {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1045.981249] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a7:e1:c0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '39cd75b0-9ec7-48ed-b57f-34da0c573a60', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c6ede9da-b209-4fae-b75a-c2f1ede4dcaf', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1045.991472] env[62385]: DEBUG oslo.service.loopingcall [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1045.992783] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1045.994018] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-977d83a4-218c-4334-a842-48ae7adeff12 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.014262] env[62385]: DEBUG oslo_vmware.api [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for the task: (returnval){ [ 1046.014262] env[62385]: value = "task-1206409" [ 1046.014262] env[62385]: _type = "Task" [ 1046.014262] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.020310] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1046.020310] env[62385]: value = "task-1206410" [ 1046.020310] env[62385]: _type = "Task" [ 1046.020310] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.027306] env[62385]: DEBUG oslo_vmware.api [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206409, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.033525] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206410, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.104949] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206408, 'name': ReconfigVM_Task, 'duration_secs': 0.325194} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.105267] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Reconfigured VM instance instance-0000005b to attach disk [datastore1] volume-925fe844-0d8f-4668-ac82-328339e88cac/volume-925fe844-0d8f-4668-ac82-328339e88cac.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1046.110700] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4604644b-43db-4c54-91d9-ccf77bc9e18d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.126837] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1046.126837] env[62385]: value = "task-1206411" [ 1046.126837] env[62385]: _type = "Task" [ 1046.126837] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.135961] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206411, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.528751] env[62385]: DEBUG oslo_vmware.api [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206409, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.537510] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206410, 'name': CreateVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.638023] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206411, 'name': ReconfigVM_Task, 'duration_secs': 0.314714} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.639732] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261292', 'volume_id': '925fe844-0d8f-4668-ac82-328339e88cac', 'name': 'volume-925fe844-0d8f-4668-ac82-328339e88cac', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '8e1d416c-2624-468c-94d7-6265e9f4178f', 'attached_at': '', 'detached_at': '', 'volume_id': '925fe844-0d8f-4668-ac82-328339e88cac', 'serial': '925fe844-0d8f-4668-ac82-328339e88cac'} {{(pid=62385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1046.640381] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-67b95d95-d3ea-4a5a-bd1c-e7aea65b57e4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.647248] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1046.647248] env[62385]: value = "task-1206412" [ 1046.647248] env[62385]: _type = "Task" [ 1046.647248] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.659024] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206412, 'name': Rename_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.696962] env[62385]: DEBUG nova.compute.manager [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1046.725743] env[62385]: DEBUG nova.virt.hardware [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1046.725990] env[62385]: DEBUG nova.virt.hardware [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1046.726169] env[62385]: DEBUG nova.virt.hardware [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1046.726360] env[62385]: DEBUG nova.virt.hardware [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1046.726515] env[62385]: DEBUG nova.virt.hardware [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1046.726665] env[62385]: DEBUG nova.virt.hardware [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1046.726878] env[62385]: DEBUG nova.virt.hardware [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1046.727056] env[62385]: DEBUG nova.virt.hardware [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1046.727241] env[62385]: DEBUG nova.virt.hardware [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1046.727407] env[62385]: DEBUG nova.virt.hardware [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1046.727585] env[62385]: DEBUG nova.virt.hardware [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1046.728497] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7bc42d1-5f8c-4f94-a05e-5974a3052273 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.738787] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2df26f0c-8001-432e-ba42-e68f89d9c303 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.845157] env[62385]: DEBUG nova.network.neutron [req-0c8b2307-de0e-4e29-930b-e06f1a97a83c req-92de5191-da96-4de3-ab10-6ef083724438 service nova] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Updated VIF entry in instance network info cache for port c6ede9da-b209-4fae-b75a-c2f1ede4dcaf. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1046.845629] env[62385]: DEBUG nova.network.neutron [req-0c8b2307-de0e-4e29-930b-e06f1a97a83c req-92de5191-da96-4de3-ab10-6ef083724438 service nova] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Updating instance_info_cache with network_info: [{"id": "c6ede9da-b209-4fae-b75a-c2f1ede4dcaf", "address": "fa:16:3e:a7:e1:c0", "network": {"id": "06017161-984c-4657-9639-4e6bdb6d4c09", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1072882119-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d45b5346eab44707b4d81eaaf12b0935", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39cd75b0-9ec7-48ed-b57f-34da0c573a60", "external-id": "nsx-vlan-transportzone-751", "segmentation_id": 751, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6ede9da-b2", "ovs_interfaceid": "c6ede9da-b209-4fae-b75a-c2f1ede4dcaf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.027957] env[62385]: DEBUG oslo_vmware.api [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206409, 'name': ReconfigVM_Task, 'duration_secs': 0.580186} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.028835] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Reconfigured VM instance instance-0000006b to attach disk [datastore1] c49d4366-2ea0-44f2-8463-daf62bd1c40c/c49d4366-2ea0-44f2-8463-daf62bd1c40c.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1047.030176] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f9b4c70e-ab3f-4413-af57-6cd38f0c1dad {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.034659] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206410, 'name': CreateVM_Task, 'duration_secs': 0.612998} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.038553] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1047.039333] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.039578] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.039984] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1047.040347] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-81ed3876-3c97-4c12-909f-fe62287d3395 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.044298] env[62385]: DEBUG oslo_vmware.api [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for the task: (returnval){ [ 1047.044298] env[62385]: value = "task-1206413" [ 1047.044298] env[62385]: _type = "Task" [ 1047.044298] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.049669] env[62385]: DEBUG oslo_vmware.api [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1047.049669] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52463072-613e-7d5b-7b1d-a09d721aee34" [ 1047.049669] env[62385]: _type = "Task" [ 1047.049669] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.056868] env[62385]: DEBUG oslo_vmware.api [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206413, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.063162] env[62385]: DEBUG oslo_vmware.api [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52463072-613e-7d5b-7b1d-a09d721aee34, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.157811] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206412, 'name': Rename_Task, 'duration_secs': 0.1588} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.158103] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1047.158375] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-85ba609f-9bfd-45c2-8914-32b1bafbc2bc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.166165] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1047.166165] env[62385]: value = "task-1206414" [ 1047.166165] env[62385]: _type = "Task" [ 1047.166165] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.174378] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206414, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.206414] env[62385]: DEBUG nova.compute.manager [req-cf761b86-8c14-47fc-a6d3-5dafe310c576 req-5d64ac2e-32ae-45f8-864f-ae55bb3cbd2a service nova] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Received event network-vif-plugged-b5615be2-bc51-4fe3-8905-4b7083550949 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1047.206550] env[62385]: DEBUG oslo_concurrency.lockutils [req-cf761b86-8c14-47fc-a6d3-5dafe310c576 req-5d64ac2e-32ae-45f8-864f-ae55bb3cbd2a service nova] Acquiring lock "5c9284ba-9f83-454f-b91b-9b0ea9ee3195-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.206698] env[62385]: DEBUG oslo_concurrency.lockutils [req-cf761b86-8c14-47fc-a6d3-5dafe310c576 req-5d64ac2e-32ae-45f8-864f-ae55bb3cbd2a service nova] Lock "5c9284ba-9f83-454f-b91b-9b0ea9ee3195-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.206870] env[62385]: DEBUG oslo_concurrency.lockutils [req-cf761b86-8c14-47fc-a6d3-5dafe310c576 req-5d64ac2e-32ae-45f8-864f-ae55bb3cbd2a service nova] Lock "5c9284ba-9f83-454f-b91b-9b0ea9ee3195-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.207264] env[62385]: DEBUG nova.compute.manager [req-cf761b86-8c14-47fc-a6d3-5dafe310c576 req-5d64ac2e-32ae-45f8-864f-ae55bb3cbd2a service nova] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] No waiting events found dispatching network-vif-plugged-b5615be2-bc51-4fe3-8905-4b7083550949 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1047.207494] env[62385]: WARNING nova.compute.manager [req-cf761b86-8c14-47fc-a6d3-5dafe310c576 req-5d64ac2e-32ae-45f8-864f-ae55bb3cbd2a service nova] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Received unexpected event network-vif-plugged-b5615be2-bc51-4fe3-8905-4b7083550949 for instance with vm_state building and task_state spawning. [ 1047.298393] env[62385]: DEBUG nova.network.neutron [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Successfully updated port: b5615be2-bc51-4fe3-8905-4b7083550949 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1047.348585] env[62385]: DEBUG oslo_concurrency.lockutils [req-0c8b2307-de0e-4e29-930b-e06f1a97a83c req-92de5191-da96-4de3-ab10-6ef083724438 service nova] Releasing lock "refresh_cache-1289e63a-5a9d-4615-b6cf-bbf5f4ae0317" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.357385] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ef8abdd8-067f-4f43-a4b0-5a7faba8a335 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Acquiring lock "a1c71d8b-0953-4cc2-961d-0c2b2fc11110" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.357646] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ef8abdd8-067f-4f43-a4b0-5a7faba8a335 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "a1c71d8b-0953-4cc2-961d-0c2b2fc11110" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.357918] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ef8abdd8-067f-4f43-a4b0-5a7faba8a335 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Acquiring lock "a1c71d8b-0953-4cc2-961d-0c2b2fc11110-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.358072] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ef8abdd8-067f-4f43-a4b0-5a7faba8a335 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "a1c71d8b-0953-4cc2-961d-0c2b2fc11110-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.358259] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ef8abdd8-067f-4f43-a4b0-5a7faba8a335 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "a1c71d8b-0953-4cc2-961d-0c2b2fc11110-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.360227] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1a2df53b-9b40-495e-badb-964f80f0803a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "interface-57637d6e-39e2-405c-bbcf-64bf0ffbebd1-331b6188-b103-437e-8436-5259bc356e31" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.360500] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1a2df53b-9b40-495e-badb-964f80f0803a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "interface-57637d6e-39e2-405c-bbcf-64bf0ffbebd1-331b6188-b103-437e-8436-5259bc356e31" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.364440] env[62385]: INFO nova.compute.manager [None req-ef8abdd8-067f-4f43-a4b0-5a7faba8a335 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Terminating instance [ 1047.367302] env[62385]: DEBUG nova.compute.manager [None req-ef8abdd8-067f-4f43-a4b0-5a7faba8a335 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1047.367521] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ef8abdd8-067f-4f43-a4b0-5a7faba8a335 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1047.368386] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d08d9d6-1956-4ced-b4db-4341417acca0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.377956] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef8abdd8-067f-4f43-a4b0-5a7faba8a335 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1047.378302] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-140f2978-4a38-49e1-ac56-1847be3c71be {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.386140] env[62385]: DEBUG oslo_vmware.api [None req-ef8abdd8-067f-4f43-a4b0-5a7faba8a335 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for the task: (returnval){ [ 1047.386140] env[62385]: value = "task-1206415" [ 1047.386140] env[62385]: _type = "Task" [ 1047.386140] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.400023] env[62385]: DEBUG oslo_vmware.api [None req-ef8abdd8-067f-4f43-a4b0-5a7faba8a335 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206415, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.557800] env[62385]: DEBUG oslo_vmware.api [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206413, 'name': Rename_Task, 'duration_secs': 0.167117} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.558591] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1047.558973] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f05042ac-99ff-466f-ac7c-cf19c3ac5d89 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.564741] env[62385]: DEBUG oslo_vmware.api [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52463072-613e-7d5b-7b1d-a09d721aee34, 'name': SearchDatastore_Task, 'duration_secs': 0.019376} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.565572] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.565932] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1047.566302] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.566606] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.566905] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1047.567276] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4e7eb362-71ed-410a-bc83-3b7317dd4a70 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.571422] env[62385]: DEBUG oslo_vmware.api [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for the task: (returnval){ [ 1047.571422] env[62385]: value = "task-1206416" [ 1047.571422] env[62385]: _type = "Task" [ 1047.571422] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.580302] env[62385]: DEBUG oslo_vmware.api [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206416, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.589198] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1047.589198] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1047.589924] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-38b19bcd-7bb8-49ef-b476-cf8da1dd920b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.596410] env[62385]: DEBUG oslo_vmware.api [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1047.596410] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5247944e-0f7f-77e2-b2c7-1b07799fda66" [ 1047.596410] env[62385]: _type = "Task" [ 1047.596410] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.607157] env[62385]: DEBUG oslo_vmware.api [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5247944e-0f7f-77e2-b2c7-1b07799fda66, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.676888] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206414, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.805698] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "refresh_cache-5c9284ba-9f83-454f-b91b-9b0ea9ee3195" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.805883] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquired lock "refresh_cache-5c9284ba-9f83-454f-b91b-9b0ea9ee3195" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.806055] env[62385]: DEBUG nova.network.neutron [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1047.866962] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1a2df53b-9b40-495e-badb-964f80f0803a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "57637d6e-39e2-405c-bbcf-64bf0ffbebd1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.867201] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1a2df53b-9b40-495e-badb-964f80f0803a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquired lock "57637d6e-39e2-405c-bbcf-64bf0ffbebd1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.868110] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4204b77-9607-4d8d-bec4-cc4c26c613c5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.888148] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3f632bf-d83b-4fc1-889a-ab1039ac45b7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.900303] env[62385]: DEBUG oslo_vmware.api [None req-ef8abdd8-067f-4f43-a4b0-5a7faba8a335 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206415, 'name': PowerOffVM_Task, 'duration_secs': 0.214501} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.916242] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef8abdd8-067f-4f43-a4b0-5a7faba8a335 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1047.916495] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ef8abdd8-067f-4f43-a4b0-5a7faba8a335 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1047.923281] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-1a2df53b-9b40-495e-badb-964f80f0803a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Reconfiguring VM to detach interface {{(pid=62385) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 1047.923607] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2b886e42-48b6-487e-b729-4b680e48e553 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.925193] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-96e99648-a211-42bb-8f0b-3e8b0623e187 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.945104] env[62385]: DEBUG oslo_vmware.api [None req-1a2df53b-9b40-495e-badb-964f80f0803a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 1047.945104] env[62385]: value = "task-1206418" [ 1047.945104] env[62385]: _type = "Task" [ 1047.945104] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.954436] env[62385]: DEBUG oslo_vmware.api [None req-1a2df53b-9b40-495e-badb-964f80f0803a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206418, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.014427] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ef8abdd8-067f-4f43-a4b0-5a7faba8a335 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1048.014705] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ef8abdd8-067f-4f43-a4b0-5a7faba8a335 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Deleting contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1048.014934] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ef8abdd8-067f-4f43-a4b0-5a7faba8a335 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Deleting the datastore file [datastore2] a1c71d8b-0953-4cc2-961d-0c2b2fc11110 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1048.015236] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fe544746-dd54-4e8f-b7fc-564cc45c8501 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.022434] env[62385]: DEBUG oslo_vmware.api [None req-ef8abdd8-067f-4f43-a4b0-5a7faba8a335 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for the task: (returnval){ [ 1048.022434] env[62385]: value = "task-1206419" [ 1048.022434] env[62385]: _type = "Task" [ 1048.022434] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.035577] env[62385]: DEBUG oslo_vmware.api [None req-ef8abdd8-067f-4f43-a4b0-5a7faba8a335 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206419, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.083320] env[62385]: DEBUG oslo_vmware.api [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206416, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.107787] env[62385]: DEBUG oslo_vmware.api [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5247944e-0f7f-77e2-b2c7-1b07799fda66, 'name': SearchDatastore_Task, 'duration_secs': 0.01384} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.108643] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca04d7e3-38c8-4a18-8b48-7d25462fc34f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.115077] env[62385]: DEBUG oslo_vmware.api [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1048.115077] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5213daf2-da83-4f1a-9628-d2d7b73d91f5" [ 1048.115077] env[62385]: _type = "Task" [ 1048.115077] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.124583] env[62385]: DEBUG oslo_vmware.api [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5213daf2-da83-4f1a-9628-d2d7b73d91f5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.177827] env[62385]: DEBUG oslo_vmware.api [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206414, 'name': PowerOnVM_Task, 'duration_secs': 0.607592} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.178153] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1048.301856] env[62385]: DEBUG nova.compute.manager [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1048.302901] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c4291bb-29ae-44e6-aad6-75d14aaad698 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.355956] env[62385]: DEBUG nova.network.neutron [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1048.455951] env[62385]: DEBUG oslo_vmware.api [None req-1a2df53b-9b40-495e-badb-964f80f0803a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206418, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.516884] env[62385]: DEBUG nova.network.neutron [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Updating instance_info_cache with network_info: [{"id": "b5615be2-bc51-4fe3-8905-4b7083550949", "address": "fa:16:3e:dd:76:b3", "network": {"id": "afae6b07-d08c-413f-aa73-b545db1bb766", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1654965885-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "93111f5ba17c47de9691f90b99e03aa8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5615be2-bc", "ovs_interfaceid": "b5615be2-bc51-4fe3-8905-4b7083550949", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1048.533862] env[62385]: DEBUG oslo_vmware.api [None req-ef8abdd8-067f-4f43-a4b0-5a7faba8a335 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Task: {'id': task-1206419, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.197973} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.534518] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ef8abdd8-067f-4f43-a4b0-5a7faba8a335 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1048.534836] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ef8abdd8-067f-4f43-a4b0-5a7faba8a335 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Deleted contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1048.535164] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ef8abdd8-067f-4f43-a4b0-5a7faba8a335 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1048.535657] env[62385]: INFO nova.compute.manager [None req-ef8abdd8-067f-4f43-a4b0-5a7faba8a335 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1048.536034] env[62385]: DEBUG oslo.service.loopingcall [None req-ef8abdd8-067f-4f43-a4b0-5a7faba8a335 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1048.536347] env[62385]: DEBUG nova.compute.manager [-] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1048.536529] env[62385]: DEBUG nova.network.neutron [-] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1048.583485] env[62385]: DEBUG oslo_vmware.api [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206416, 'name': PowerOnVM_Task, 'duration_secs': 0.697319} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.583900] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1048.584283] env[62385]: INFO nova.compute.manager [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Took 9.97 seconds to spawn the instance on the hypervisor. [ 1048.584589] env[62385]: DEBUG nova.compute.manager [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1048.585531] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f043ca2-f969-45b4-bb26-5ac48eee7d2d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.635027] env[62385]: DEBUG oslo_vmware.api [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5213daf2-da83-4f1a-9628-d2d7b73d91f5, 'name': SearchDatastore_Task, 'duration_secs': 0.010594} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.635027] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.635027] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317/1289e63a-5a9d-4615-b6cf-bbf5f4ae0317.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1048.635247] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c4458f4c-54cd-4d63-aaf7-83c16cad4d1d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.648457] env[62385]: DEBUG oslo_vmware.api [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1048.648457] env[62385]: value = "task-1206420" [ 1048.648457] env[62385]: _type = "Task" [ 1048.648457] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.656940] env[62385]: DEBUG oslo_vmware.api [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206420, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.826139] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a2b9f85c-a335-46c3-bdbb-634bd231934d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lock "8e1d416c-2624-468c-94d7-6265e9f4178f" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 35.709s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.957878] env[62385]: DEBUG oslo_vmware.api [None req-1a2df53b-9b40-495e-badb-964f80f0803a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206418, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.018366] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Releasing lock "refresh_cache-5c9284ba-9f83-454f-b91b-9b0ea9ee3195" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1049.018366] env[62385]: DEBUG nova.compute.manager [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Instance network_info: |[{"id": "b5615be2-bc51-4fe3-8905-4b7083550949", "address": "fa:16:3e:dd:76:b3", "network": {"id": "afae6b07-d08c-413f-aa73-b545db1bb766", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1654965885-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "93111f5ba17c47de9691f90b99e03aa8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5615be2-bc", "ovs_interfaceid": "b5615be2-bc51-4fe3-8905-4b7083550949", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1049.018758] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:dd:76:b3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4adc8ed0-d11a-4510-9be0-b27c0da3a903', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b5615be2-bc51-4fe3-8905-4b7083550949', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1049.028349] env[62385]: DEBUG oslo.service.loopingcall [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1049.028604] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1049.028838] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6b0b6ab1-a9dd-480e-a0b0-4b8600b39382 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.053527] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1049.053527] env[62385]: value = "task-1206421" [ 1049.053527] env[62385]: _type = "Task" [ 1049.053527] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.063277] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206421, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.112190] env[62385]: INFO nova.compute.manager [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Took 20.72 seconds to build instance. [ 1049.159926] env[62385]: DEBUG oslo_vmware.api [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206420, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.255350] env[62385]: DEBUG nova.compute.manager [req-c96e8e20-476a-408c-af49-970950964eb8 req-3446b88a-a8e8-4f2f-ba3f-e02b8b05c209 service nova] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Received event network-changed-b5615be2-bc51-4fe3-8905-4b7083550949 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1049.255350] env[62385]: DEBUG nova.compute.manager [req-c96e8e20-476a-408c-af49-970950964eb8 req-3446b88a-a8e8-4f2f-ba3f-e02b8b05c209 service nova] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Refreshing instance network info cache due to event network-changed-b5615be2-bc51-4fe3-8905-4b7083550949. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1049.255350] env[62385]: DEBUG oslo_concurrency.lockutils [req-c96e8e20-476a-408c-af49-970950964eb8 req-3446b88a-a8e8-4f2f-ba3f-e02b8b05c209 service nova] Acquiring lock "refresh_cache-5c9284ba-9f83-454f-b91b-9b0ea9ee3195" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1049.255350] env[62385]: DEBUG oslo_concurrency.lockutils [req-c96e8e20-476a-408c-af49-970950964eb8 req-3446b88a-a8e8-4f2f-ba3f-e02b8b05c209 service nova] Acquired lock "refresh_cache-5c9284ba-9f83-454f-b91b-9b0ea9ee3195" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1049.255350] env[62385]: DEBUG nova.network.neutron [req-c96e8e20-476a-408c-af49-970950964eb8 req-3446b88a-a8e8-4f2f-ba3f-e02b8b05c209 service nova] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Refreshing network info cache for port b5615be2-bc51-4fe3-8905-4b7083550949 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1049.457859] env[62385]: DEBUG oslo_vmware.api [None req-1a2df53b-9b40-495e-badb-964f80f0803a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206418, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.517163] env[62385]: DEBUG nova.network.neutron [-] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.566615] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206421, 'name': CreateVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.615765] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b292a0a0-1d4a-4f4e-9eb4-a8f5fb15df72 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Lock "c49d4366-2ea0-44f2-8463-daf62bd1c40c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.239s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.662232] env[62385]: DEBUG oslo_vmware.api [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206420, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.842478} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.662232] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317/1289e63a-5a9d-4615-b6cf-bbf5f4ae0317.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1049.662232] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1049.662232] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-493bba24-ce75-4d07-a449-e60c38ec6a26 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.671210] env[62385]: DEBUG oslo_vmware.api [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1049.671210] env[62385]: value = "task-1206422" [ 1049.671210] env[62385]: _type = "Task" [ 1049.671210] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.681589] env[62385]: DEBUG oslo_vmware.api [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206422, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.957730] env[62385]: DEBUG oslo_vmware.api [None req-1a2df53b-9b40-495e-badb-964f80f0803a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206418, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.978409] env[62385]: DEBUG nova.network.neutron [req-c96e8e20-476a-408c-af49-970950964eb8 req-3446b88a-a8e8-4f2f-ba3f-e02b8b05c209 service nova] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Updated VIF entry in instance network info cache for port b5615be2-bc51-4fe3-8905-4b7083550949. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1049.978770] env[62385]: DEBUG nova.network.neutron [req-c96e8e20-476a-408c-af49-970950964eb8 req-3446b88a-a8e8-4f2f-ba3f-e02b8b05c209 service nova] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Updating instance_info_cache with network_info: [{"id": "b5615be2-bc51-4fe3-8905-4b7083550949", "address": "fa:16:3e:dd:76:b3", "network": {"id": "afae6b07-d08c-413f-aa73-b545db1bb766", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1654965885-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "93111f5ba17c47de9691f90b99e03aa8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5615be2-bc", "ovs_interfaceid": "b5615be2-bc51-4fe3-8905-4b7083550949", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1050.022575] env[62385]: INFO nova.compute.manager [-] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Took 1.49 seconds to deallocate network for instance. [ 1050.065821] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206421, 'name': CreateVM_Task, 'duration_secs': 0.630501} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.066109] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1050.067136] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1050.067356] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.067661] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1050.067932] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-50b19856-a977-4e45-8e3a-c3607f6019d0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.073235] env[62385]: DEBUG oslo_vmware.api [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1050.073235] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52dee351-e44d-cc98-fe7e-a233d4dacf68" [ 1050.073235] env[62385]: _type = "Task" [ 1050.073235] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.083949] env[62385]: DEBUG oslo_vmware.api [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52dee351-e44d-cc98-fe7e-a233d4dacf68, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.181984] env[62385]: DEBUG oslo_vmware.api [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206422, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.200113} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.182341] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1050.183184] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66f677f9-fba6-4627-9aab-36c4691ea827 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.208046] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317/1289e63a-5a9d-4615-b6cf-bbf5f4ae0317.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1050.208392] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7bb15613-e169-41bd-b804-8a044611ab61 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.233020] env[62385]: DEBUG oslo_vmware.api [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1050.233020] env[62385]: value = "task-1206423" [ 1050.233020] env[62385]: _type = "Task" [ 1050.233020] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.247172] env[62385]: DEBUG oslo_vmware.api [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206423, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.459168] env[62385]: DEBUG oslo_vmware.api [None req-1a2df53b-9b40-495e-badb-964f80f0803a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206418, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.482882] env[62385]: DEBUG oslo_concurrency.lockutils [req-c96e8e20-476a-408c-af49-970950964eb8 req-3446b88a-a8e8-4f2f-ba3f-e02b8b05c209 service nova] Releasing lock "refresh_cache-5c9284ba-9f83-454f-b91b-9b0ea9ee3195" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1050.482882] env[62385]: DEBUG nova.compute.manager [req-c96e8e20-476a-408c-af49-970950964eb8 req-3446b88a-a8e8-4f2f-ba3f-e02b8b05c209 service nova] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Received event network-vif-deleted-20926145-a8ac-4f6f-afb1-a1c449d1d76d {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1050.482882] env[62385]: INFO nova.compute.manager [req-c96e8e20-476a-408c-af49-970950964eb8 req-3446b88a-a8e8-4f2f-ba3f-e02b8b05c209 service nova] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Neutron deleted interface 20926145-a8ac-4f6f-afb1-a1c449d1d76d; detaching it from the instance and deleting it from the info cache [ 1050.482882] env[62385]: DEBUG nova.network.neutron [req-c96e8e20-476a-408c-af49-970950964eb8 req-3446b88a-a8e8-4f2f-ba3f-e02b8b05c209 service nova] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1050.532291] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ef8abdd8-067f-4f43-a4b0-5a7faba8a335 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.532584] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ef8abdd8-067f-4f43-a4b0-5a7faba8a335 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.532814] env[62385]: DEBUG nova.objects.instance [None req-ef8abdd8-067f-4f43-a4b0-5a7faba8a335 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lazy-loading 'resources' on Instance uuid a1c71d8b-0953-4cc2-961d-0c2b2fc11110 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1050.586208] env[62385]: DEBUG oslo_vmware.api [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52dee351-e44d-cc98-fe7e-a233d4dacf68, 'name': SearchDatastore_Task, 'duration_secs': 0.054399} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.589029] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1050.589904] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1050.590616] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1050.590616] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.593030] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1050.593401] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4ef89950-8f2b-43f8-9add-6958d221e7a4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.607058] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1050.608381] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1050.609277] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3b5ffc67-0c9e-4dae-be46-150710fdd5f3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.615784] env[62385]: DEBUG oslo_vmware.api [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1050.615784] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521257b0-71ac-e08a-ef6e-b9da7ec4b6c0" [ 1050.615784] env[62385]: _type = "Task" [ 1050.615784] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.624713] env[62385]: DEBUG oslo_vmware.api [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521257b0-71ac-e08a-ef6e-b9da7ec4b6c0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.743555] env[62385]: DEBUG oslo_vmware.api [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206423, 'name': ReconfigVM_Task, 'duration_secs': 0.465914} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.743944] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Reconfigured VM instance instance-0000006c to attach disk [datastore1] 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317/1289e63a-5a9d-4615-b6cf-bbf5f4ae0317.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1050.744649] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3ad3b28a-ba68-48f2-a80c-15b6c3d05240 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.752723] env[62385]: DEBUG oslo_vmware.api [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1050.752723] env[62385]: value = "task-1206424" [ 1050.752723] env[62385]: _type = "Task" [ 1050.752723] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.764448] env[62385]: DEBUG oslo_vmware.api [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206424, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.959611] env[62385]: DEBUG oslo_vmware.api [None req-1a2df53b-9b40-495e-badb-964f80f0803a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206418, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.987659] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-504d5e6f-ce25-4ebb-9b5f-7e3026b0bf76 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.998491] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a9a959a-ef57-4d3a-9c6a-ade63a9594ea {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.034149] env[62385]: DEBUG nova.compute.manager [req-c96e8e20-476a-408c-af49-970950964eb8 req-3446b88a-a8e8-4f2f-ba3f-e02b8b05c209 service nova] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Detach interface failed, port_id=20926145-a8ac-4f6f-afb1-a1c449d1d76d, reason: Instance a1c71d8b-0953-4cc2-961d-0c2b2fc11110 could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1051.108644] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Acquiring lock "f93d9f5d-f7b6-4988-af7c-877a9e4c2104" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.108983] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Lock "f93d9f5d-f7b6-4988-af7c-877a9e4c2104" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.130280] env[62385]: DEBUG oslo_vmware.api [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521257b0-71ac-e08a-ef6e-b9da7ec4b6c0, 'name': SearchDatastore_Task, 'duration_secs': 0.010521} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.131196] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4cdfb556-6c7c-43e0-ab17-a265d4125360 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.141702] env[62385]: DEBUG oslo_vmware.api [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1051.141702] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52214632-0383-0310-83a6-70334bd98fdf" [ 1051.141702] env[62385]: _type = "Task" [ 1051.141702] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.152718] env[62385]: DEBUG oslo_vmware.api [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52214632-0383-0310-83a6-70334bd98fdf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.238029] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-663cf90e-9db9-4a7a-935d-517f9a97153c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.249131] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61bfc8db-0729-4ede-88e1-ba5c4082e9f7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.970426] env[62385]: DEBUG nova.compute.manager [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1052.005351] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5e62971-431e-44b6-b205-da412f74e5d1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.008682] env[62385]: DEBUG oslo_vmware.api [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206424, 'name': Rename_Task, 'duration_secs': 0.17838} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.009341] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1052.009603] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2a5696d9-46ef-4882-8a3c-57a7fe3cf355 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.019084] env[62385]: DEBUG oslo_vmware.api [None req-1a2df53b-9b40-495e-badb-964f80f0803a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206418, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.019362] env[62385]: DEBUG oslo_vmware.api [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52214632-0383-0310-83a6-70334bd98fdf, 'name': SearchDatastore_Task, 'duration_secs': 0.033833} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.020056] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1052.020297] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 5c9284ba-9f83-454f-b91b-9b0ea9ee3195/5c9284ba-9f83-454f-b91b-9b0ea9ee3195.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1052.022108] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-126304ba-f622-443a-bd52-b2024d582d32 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.026872] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bd75253c-806a-4d31-8201-6d17ed11dab1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.028987] env[62385]: DEBUG oslo_vmware.api [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1052.028987] env[62385]: value = "task-1206425" [ 1052.028987] env[62385]: _type = "Task" [ 1052.028987] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.040986] env[62385]: DEBUG nova.compute.provider_tree [None req-ef8abdd8-067f-4f43-a4b0-5a7faba8a335 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1052.043983] env[62385]: DEBUG oslo_vmware.api [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1052.043983] env[62385]: value = "task-1206426" [ 1052.043983] env[62385]: _type = "Task" [ 1052.043983] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.050994] env[62385]: DEBUG oslo_vmware.api [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206425, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.057137] env[62385]: DEBUG oslo_vmware.api [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206426, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.482481] env[62385]: DEBUG oslo_vmware.api [None req-1a2df53b-9b40-495e-badb-964f80f0803a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206418, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.532608] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.541731] env[62385]: DEBUG oslo_vmware.api [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206425, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.545849] env[62385]: DEBUG nova.scheduler.client.report [None req-ef8abdd8-067f-4f43-a4b0-5a7faba8a335 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1052.559156] env[62385]: DEBUG oslo_vmware.api [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206426, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.982677] env[62385]: DEBUG oslo_vmware.api [None req-1a2df53b-9b40-495e-badb-964f80f0803a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206418, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.987134] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Acquiring lock "2cb0fee4-c922-402a-9251-10dc0da8c87d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.987361] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Lock "2cb0fee4-c922-402a-9251-10dc0da8c87d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.042795] env[62385]: DEBUG oslo_vmware.api [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206425, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.054482] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ef8abdd8-067f-4f43-a4b0-5a7faba8a335 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.522s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.059522] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.528s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.064539] env[62385]: INFO nova.compute.claims [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1053.067422] env[62385]: DEBUG oslo_vmware.api [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206426, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.587014} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.068603] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 5c9284ba-9f83-454f-b91b-9b0ea9ee3195/5c9284ba-9f83-454f-b91b-9b0ea9ee3195.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1053.068933] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1053.069297] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8f43907d-e251-4498-ad8c-9b04da01be88 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.078709] env[62385]: DEBUG oslo_vmware.api [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1053.078709] env[62385]: value = "task-1206427" [ 1053.078709] env[62385]: _type = "Task" [ 1053.078709] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.085553] env[62385]: INFO nova.scheduler.client.report [None req-ef8abdd8-067f-4f43-a4b0-5a7faba8a335 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Deleted allocations for instance a1c71d8b-0953-4cc2-961d-0c2b2fc11110 [ 1053.100062] env[62385]: DEBUG oslo_vmware.api [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206427, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.482744] env[62385]: DEBUG oslo_vmware.api [None req-1a2df53b-9b40-495e-badb-964f80f0803a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206418, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.489348] env[62385]: DEBUG nova.compute.manager [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1053.542891] env[62385]: DEBUG oslo_vmware.api [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206425, 'name': PowerOnVM_Task, 'duration_secs': 1.124471} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.543185] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1053.543391] env[62385]: INFO nova.compute.manager [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Took 10.88 seconds to spawn the instance on the hypervisor. [ 1053.543575] env[62385]: DEBUG nova.compute.manager [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1053.544427] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06a8e550-774a-45fb-a3e5-1e0898ba5553 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.590853] env[62385]: DEBUG oslo_vmware.api [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206427, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071918} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.591824] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1053.594996] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feab7367-6438-435a-a563-737f44e53a96 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.597889] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ef8abdd8-067f-4f43-a4b0-5a7faba8a335 tempest-AttachVolumeTestJSON-1167979113 tempest-AttachVolumeTestJSON-1167979113-project-member] Lock "a1c71d8b-0953-4cc2-961d-0c2b2fc11110" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.240s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.621545] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] 5c9284ba-9f83-454f-b91b-9b0ea9ee3195/5c9284ba-9f83-454f-b91b-9b0ea9ee3195.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1053.622602] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d4d8d68d-dc4b-4406-84d4-fccb8c979088 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.645611] env[62385]: DEBUG oslo_vmware.api [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1053.645611] env[62385]: value = "task-1206428" [ 1053.645611] env[62385]: _type = "Task" [ 1053.645611] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.655166] env[62385]: DEBUG oslo_vmware.api [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206428, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.982891] env[62385]: DEBUG oslo_vmware.api [None req-1a2df53b-9b40-495e-badb-964f80f0803a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206418, 'name': ReconfigVM_Task, 'duration_secs': 5.879132} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.982891] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1a2df53b-9b40-495e-badb-964f80f0803a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Releasing lock "57637d6e-39e2-405c-bbcf-64bf0ffbebd1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1053.982891] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-1a2df53b-9b40-495e-badb-964f80f0803a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Reconfigured VM to detach interface {{(pid=62385) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 1054.011143] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.064110] env[62385]: INFO nova.compute.manager [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Took 16.32 seconds to build instance. [ 1054.158376] env[62385]: DEBUG oslo_vmware.api [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206428, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.219881] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f33f1dcf-1a3a-422b-a61b-2bb4ae586270 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.227550] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6cee7e1-4b5e-4a69-ab41-88e967d870e8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.264013] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f890ab5c-cd32-41ae-a34f-45106539d121 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.271544] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-413d0a0f-19cb-4851-9d73-72521eecbf9f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.286186] env[62385]: DEBUG nova.compute.provider_tree [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1054.566661] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b963a4dc-3b69-422a-be54-7ece0afd57e5 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "1289e63a-5a9d-4615-b6cf-bbf5f4ae0317" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.840s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.655645] env[62385]: DEBUG oslo_vmware.api [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206428, 'name': ReconfigVM_Task, 'duration_secs': 0.660618} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.655928] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Reconfigured VM instance instance-0000006d to attach disk [datastore2] 5c9284ba-9f83-454f-b91b-9b0ea9ee3195/5c9284ba-9f83-454f-b91b-9b0ea9ee3195.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1054.656562] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4e71941c-14ae-413e-9499-dc93dc4ca5e2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.662813] env[62385]: DEBUG oslo_vmware.api [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1054.662813] env[62385]: value = "task-1206430" [ 1054.662813] env[62385]: _type = "Task" [ 1054.662813] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.671860] env[62385]: DEBUG oslo_vmware.api [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206430, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.789931] env[62385]: DEBUG nova.scheduler.client.report [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1055.173394] env[62385]: DEBUG oslo_vmware.api [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206430, 'name': Rename_Task, 'duration_secs': 0.157959} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.173694] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1055.173953] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2a1bbec1-4789-4d84-8b80-42529d168e7f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.180774] env[62385]: DEBUG oslo_vmware.api [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1055.180774] env[62385]: value = "task-1206431" [ 1055.180774] env[62385]: _type = "Task" [ 1055.180774] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.188364] env[62385]: DEBUG oslo_vmware.api [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206431, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.299078] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.239s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.299636] env[62385]: DEBUG nova.compute.manager [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1055.302384] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.291s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.307296] env[62385]: INFO nova.compute.claims [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1055.314588] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1a2df53b-9b40-495e-badb-964f80f0803a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "refresh_cache-57637d6e-39e2-405c-bbcf-64bf0ffbebd1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1055.314674] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1a2df53b-9b40-495e-badb-964f80f0803a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquired lock "refresh_cache-57637d6e-39e2-405c-bbcf-64bf0ffbebd1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1055.314849] env[62385]: DEBUG nova.network.neutron [None req-1a2df53b-9b40-495e-badb-964f80f0803a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1055.316806] env[62385]: DEBUG nova.compute.manager [req-e56c9ea1-2e58-4e59-92d0-4982d6c59612 req-f2178658-3f75-4c2b-9937-d7729d4b110f service nova] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Received event network-changed-c6ede9da-b209-4fae-b75a-c2f1ede4dcaf {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1055.317047] env[62385]: DEBUG nova.compute.manager [req-e56c9ea1-2e58-4e59-92d0-4982d6c59612 req-f2178658-3f75-4c2b-9937-d7729d4b110f service nova] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Refreshing instance network info cache due to event network-changed-c6ede9da-b209-4fae-b75a-c2f1ede4dcaf. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1055.321325] env[62385]: DEBUG oslo_concurrency.lockutils [req-e56c9ea1-2e58-4e59-92d0-4982d6c59612 req-f2178658-3f75-4c2b-9937-d7729d4b110f service nova] Acquiring lock "refresh_cache-1289e63a-5a9d-4615-b6cf-bbf5f4ae0317" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1055.321515] env[62385]: DEBUG oslo_concurrency.lockutils [req-e56c9ea1-2e58-4e59-92d0-4982d6c59612 req-f2178658-3f75-4c2b-9937-d7729d4b110f service nova] Acquired lock "refresh_cache-1289e63a-5a9d-4615-b6cf-bbf5f4ae0317" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1055.321666] env[62385]: DEBUG nova.network.neutron [req-e56c9ea1-2e58-4e59-92d0-4982d6c59612 req-f2178658-3f75-4c2b-9937-d7729d4b110f service nova] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Refreshing network info cache for port c6ede9da-b209-4fae-b75a-c2f1ede4dcaf {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1055.692814] env[62385]: DEBUG oslo_vmware.api [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206431, 'name': PowerOnVM_Task, 'duration_secs': 0.494322} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.693249] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1055.693396] env[62385]: INFO nova.compute.manager [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Took 9.00 seconds to spawn the instance on the hypervisor. [ 1055.693898] env[62385]: DEBUG nova.compute.manager [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1055.694686] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0199afa2-dff7-4601-a6ae-a46022960010 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.811616] env[62385]: DEBUG nova.compute.utils [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1055.813050] env[62385]: DEBUG nova.compute.manager [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1055.817054] env[62385]: DEBUG nova.network.neutron [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1055.878797] env[62385]: DEBUG nova.policy [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b178ea9a750645bea32dcc2873304ec0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5d716d19f79945db98f684df0cf4b302', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 1056.165819] env[62385]: DEBUG nova.network.neutron [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Successfully created port: dd3797d4-87fb-4629-8b74-115e823503bb {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1056.216704] env[62385]: INFO nova.compute.manager [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Took 14.43 seconds to build instance. [ 1056.320227] env[62385]: DEBUG nova.compute.manager [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1056.350618] env[62385]: INFO nova.network.neutron [None req-1a2df53b-9b40-495e-badb-964f80f0803a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Port 331b6188-b103-437e-8436-5259bc356e31 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1056.350968] env[62385]: DEBUG nova.network.neutron [None req-1a2df53b-9b40-495e-badb-964f80f0803a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Updating instance_info_cache with network_info: [{"id": "a7153a2d-db69-4d14-8ddc-b634ddabbc18", "address": "fa:16:3e:a3:0d:ab", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7153a2d-db", "ovs_interfaceid": "a7153a2d-db69-4d14-8ddc-b634ddabbc18", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1056.372455] env[62385]: DEBUG nova.network.neutron [req-e56c9ea1-2e58-4e59-92d0-4982d6c59612 req-f2178658-3f75-4c2b-9937-d7729d4b110f service nova] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Updated VIF entry in instance network info cache for port c6ede9da-b209-4fae-b75a-c2f1ede4dcaf. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1056.373740] env[62385]: DEBUG nova.network.neutron [req-e56c9ea1-2e58-4e59-92d0-4982d6c59612 req-f2178658-3f75-4c2b-9937-d7729d4b110f service nova] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Updating instance_info_cache with network_info: [{"id": "c6ede9da-b209-4fae-b75a-c2f1ede4dcaf", "address": "fa:16:3e:a7:e1:c0", "network": {"id": "06017161-984c-4657-9639-4e6bdb6d4c09", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1072882119-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.208", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d45b5346eab44707b4d81eaaf12b0935", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39cd75b0-9ec7-48ed-b57f-34da0c573a60", "external-id": "nsx-vlan-transportzone-751", "segmentation_id": 751, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6ede9da-b2", "ovs_interfaceid": "c6ede9da-b209-4fae-b75a-c2f1ede4dcaf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1056.495939] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfff7517-690d-4b0f-a319-76124853afdc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.506538] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-614f9954-50f0-4941-8e64-dc7e40d4a483 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.542218] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74d5674d-d33a-45b0-9362-3b63b0601d12 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.549877] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d449000-eea4-45ba-8a09-0009c12606c3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.563261] env[62385]: DEBUG nova.compute.provider_tree [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1056.719974] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a2f603f4-9ec6-421b-88c1-6ba52b842954 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "5c9284ba-9f83-454f-b91b-9b0ea9ee3195" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.941s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.854247] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1a2df53b-9b40-495e-badb-964f80f0803a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Releasing lock "refresh_cache-57637d6e-39e2-405c-bbcf-64bf0ffbebd1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1056.880808] env[62385]: DEBUG oslo_concurrency.lockutils [req-e56c9ea1-2e58-4e59-92d0-4982d6c59612 req-f2178658-3f75-4c2b-9937-d7729d4b110f service nova] Releasing lock "refresh_cache-1289e63a-5a9d-4615-b6cf-bbf5f4ae0317" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1057.066420] env[62385]: DEBUG nova.scheduler.client.report [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1057.181982] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8bde1141-afeb-4c33-a510-88e0cc945e77 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "interface-68748d25-017f-4b58-9984-70264abbc5ba-331b6188-b103-437e-8436-5259bc356e31" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.182313] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8bde1141-afeb-4c33-a510-88e0cc945e77 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "interface-68748d25-017f-4b58-9984-70264abbc5ba-331b6188-b103-437e-8436-5259bc356e31" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.182679] env[62385]: DEBUG nova.objects.instance [None req-8bde1141-afeb-4c33-a510-88e0cc945e77 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lazy-loading 'flavor' on Instance uuid 68748d25-017f-4b58-9984-70264abbc5ba {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1057.334361] env[62385]: DEBUG nova.compute.manager [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1057.359359] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1a2df53b-9b40-495e-badb-964f80f0803a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "interface-57637d6e-39e2-405c-bbcf-64bf0ffbebd1-331b6188-b103-437e-8436-5259bc356e31" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.999s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.364763] env[62385]: DEBUG nova.virt.hardware [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1057.365012] env[62385]: DEBUG nova.virt.hardware [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1057.365202] env[62385]: DEBUG nova.virt.hardware [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1057.365425] env[62385]: DEBUG nova.virt.hardware [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1057.365605] env[62385]: DEBUG nova.virt.hardware [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1057.365871] env[62385]: DEBUG nova.virt.hardware [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1057.366064] env[62385]: DEBUG nova.virt.hardware [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1057.366217] env[62385]: DEBUG nova.virt.hardware [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1057.366423] env[62385]: DEBUG nova.virt.hardware [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1057.366619] env[62385]: DEBUG nova.virt.hardware [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1057.366845] env[62385]: DEBUG nova.virt.hardware [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1057.367808] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eac65196-2e32-4ab9-b5a8-f0714aae678e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.376392] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35093278-b537-4500-8ff5-bfce306bf57f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.426206] env[62385]: DEBUG nova.compute.manager [req-1aeba257-8c93-4094-aa5c-15b2c4084fdb req-2e82a79e-6d07-4667-8c4f-f0c70f30c4f1 service nova] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Received event network-changed-a7153a2d-db69-4d14-8ddc-b634ddabbc18 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1057.426399] env[62385]: DEBUG nova.compute.manager [req-1aeba257-8c93-4094-aa5c-15b2c4084fdb req-2e82a79e-6d07-4667-8c4f-f0c70f30c4f1 service nova] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Refreshing instance network info cache due to event network-changed-a7153a2d-db69-4d14-8ddc-b634ddabbc18. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1057.426619] env[62385]: DEBUG oslo_concurrency.lockutils [req-1aeba257-8c93-4094-aa5c-15b2c4084fdb req-2e82a79e-6d07-4667-8c4f-f0c70f30c4f1 service nova] Acquiring lock "refresh_cache-57637d6e-39e2-405c-bbcf-64bf0ffbebd1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.426793] env[62385]: DEBUG oslo_concurrency.lockutils [req-1aeba257-8c93-4094-aa5c-15b2c4084fdb req-2e82a79e-6d07-4667-8c4f-f0c70f30c4f1 service nova] Acquired lock "refresh_cache-57637d6e-39e2-405c-bbcf-64bf0ffbebd1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.426921] env[62385]: DEBUG nova.network.neutron [req-1aeba257-8c93-4094-aa5c-15b2c4084fdb req-2e82a79e-6d07-4667-8c4f-f0c70f30c4f1 service nova] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Refreshing network info cache for port a7153a2d-db69-4d14-8ddc-b634ddabbc18 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1057.572562] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.270s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.574449] env[62385]: DEBUG nova.compute.manager [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1057.679015] env[62385]: DEBUG nova.compute.manager [req-691ad572-ba2c-4de0-b761-6736d5c67c80 req-24fe8463-2bcc-4ae4-9425-07bff1fd3e99 service nova] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Received event network-vif-plugged-dd3797d4-87fb-4629-8b74-115e823503bb {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1057.679179] env[62385]: DEBUG oslo_concurrency.lockutils [req-691ad572-ba2c-4de0-b761-6736d5c67c80 req-24fe8463-2bcc-4ae4-9425-07bff1fd3e99 service nova] Acquiring lock "f93d9f5d-f7b6-4988-af7c-877a9e4c2104-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.679414] env[62385]: DEBUG oslo_concurrency.lockutils [req-691ad572-ba2c-4de0-b761-6736d5c67c80 req-24fe8463-2bcc-4ae4-9425-07bff1fd3e99 service nova] Lock "f93d9f5d-f7b6-4988-af7c-877a9e4c2104-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.679593] env[62385]: DEBUG oslo_concurrency.lockutils [req-691ad572-ba2c-4de0-b761-6736d5c67c80 req-24fe8463-2bcc-4ae4-9425-07bff1fd3e99 service nova] Lock "f93d9f5d-f7b6-4988-af7c-877a9e4c2104-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.679780] env[62385]: DEBUG nova.compute.manager [req-691ad572-ba2c-4de0-b761-6736d5c67c80 req-24fe8463-2bcc-4ae4-9425-07bff1fd3e99 service nova] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] No waiting events found dispatching network-vif-plugged-dd3797d4-87fb-4629-8b74-115e823503bb {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1057.679944] env[62385]: WARNING nova.compute.manager [req-691ad572-ba2c-4de0-b761-6736d5c67c80 req-24fe8463-2bcc-4ae4-9425-07bff1fd3e99 service nova] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Received unexpected event network-vif-plugged-dd3797d4-87fb-4629-8b74-115e823503bb for instance with vm_state building and task_state spawning. [ 1057.752158] env[62385]: DEBUG nova.network.neutron [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Successfully updated port: dd3797d4-87fb-4629-8b74-115e823503bb {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1057.835404] env[62385]: DEBUG nova.objects.instance [None req-8bde1141-afeb-4c33-a510-88e0cc945e77 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lazy-loading 'pci_requests' on Instance uuid 68748d25-017f-4b58-9984-70264abbc5ba {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1058.080443] env[62385]: DEBUG nova.compute.utils [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1058.081872] env[62385]: DEBUG nova.compute.manager [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1058.082058] env[62385]: DEBUG nova.network.neutron [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1058.125068] env[62385]: DEBUG nova.policy [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '749efb5d80264a77909ba1b8796f3176', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cbdb8eed888d42129200077fc2ffb17f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 1058.207523] env[62385]: DEBUG nova.network.neutron [req-1aeba257-8c93-4094-aa5c-15b2c4084fdb req-2e82a79e-6d07-4667-8c4f-f0c70f30c4f1 service nova] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Updated VIF entry in instance network info cache for port a7153a2d-db69-4d14-8ddc-b634ddabbc18. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1058.207993] env[62385]: DEBUG nova.network.neutron [req-1aeba257-8c93-4094-aa5c-15b2c4084fdb req-2e82a79e-6d07-4667-8c4f-f0c70f30c4f1 service nova] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Updating instance_info_cache with network_info: [{"id": "a7153a2d-db69-4d14-8ddc-b634ddabbc18", "address": "fa:16:3e:a3:0d:ab", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7153a2d-db", "ovs_interfaceid": "a7153a2d-db69-4d14-8ddc-b634ddabbc18", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.255758] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Acquiring lock "refresh_cache-f93d9f5d-f7b6-4988-af7c-877a9e4c2104" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.255758] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Acquired lock "refresh_cache-f93d9f5d-f7b6-4988-af7c-877a9e4c2104" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.255758] env[62385]: DEBUG nova.network.neutron [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1058.336315] env[62385]: DEBUG nova.objects.base [None req-8bde1141-afeb-4c33-a510-88e0cc945e77 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Object Instance<68748d25-017f-4b58-9984-70264abbc5ba> lazy-loaded attributes: flavor,pci_requests {{(pid=62385) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1058.336382] env[62385]: DEBUG nova.network.neutron [None req-8bde1141-afeb-4c33-a510-88e0cc945e77 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1058.396816] env[62385]: DEBUG nova.network.neutron [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Successfully created port: bc49aae7-4b0a-4cef-9b58-1baefe8a64ce {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1058.411112] env[62385]: DEBUG nova.policy [None req-8bde1141-afeb-4c33-a510-88e0cc945e77 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5040b20c848c4d28a198f1773a3ff21e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '012426f098ce40c3aaa00f628fe9cebb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 1058.587919] env[62385]: DEBUG nova.compute.manager [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1058.711369] env[62385]: DEBUG oslo_concurrency.lockutils [req-1aeba257-8c93-4094-aa5c-15b2c4084fdb req-2e82a79e-6d07-4667-8c4f-f0c70f30c4f1 service nova] Releasing lock "refresh_cache-57637d6e-39e2-405c-bbcf-64bf0ffbebd1" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.711834] env[62385]: DEBUG nova.compute.manager [req-1aeba257-8c93-4094-aa5c-15b2c4084fdb req-2e82a79e-6d07-4667-8c4f-f0c70f30c4f1 service nova] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Received event network-changed-c308701b-9b9e-4ff3-9a01-c27aabf54f0d {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1058.712137] env[62385]: DEBUG nova.compute.manager [req-1aeba257-8c93-4094-aa5c-15b2c4084fdb req-2e82a79e-6d07-4667-8c4f-f0c70f30c4f1 service nova] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Refreshing instance network info cache due to event network-changed-c308701b-9b9e-4ff3-9a01-c27aabf54f0d. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1058.712419] env[62385]: DEBUG oslo_concurrency.lockutils [req-1aeba257-8c93-4094-aa5c-15b2c4084fdb req-2e82a79e-6d07-4667-8c4f-f0c70f30c4f1 service nova] Acquiring lock "refresh_cache-68748d25-017f-4b58-9984-70264abbc5ba" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.712576] env[62385]: DEBUG oslo_concurrency.lockutils [req-1aeba257-8c93-4094-aa5c-15b2c4084fdb req-2e82a79e-6d07-4667-8c4f-f0c70f30c4f1 service nova] Acquired lock "refresh_cache-68748d25-017f-4b58-9984-70264abbc5ba" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.712751] env[62385]: DEBUG nova.network.neutron [req-1aeba257-8c93-4094-aa5c-15b2c4084fdb req-2e82a79e-6d07-4667-8c4f-f0c70f30c4f1 service nova] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Refreshing network info cache for port c308701b-9b9e-4ff3-9a01-c27aabf54f0d {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1058.793514] env[62385]: DEBUG nova.network.neutron [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1058.922089] env[62385]: DEBUG nova.network.neutron [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Updating instance_info_cache with network_info: [{"id": "dd3797d4-87fb-4629-8b74-115e823503bb", "address": "fa:16:3e:57:29:40", "network": {"id": "a561a523-c9d8-4169-b449-5c0bdef7b53b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-751953866-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d716d19f79945db98f684df0cf4b302", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd3797d4-87", "ovs_interfaceid": "dd3797d4-87fb-4629-8b74-115e823503bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1059.424830] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Releasing lock "refresh_cache-f93d9f5d-f7b6-4988-af7c-877a9e4c2104" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.425241] env[62385]: DEBUG nova.compute.manager [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Instance network_info: |[{"id": "dd3797d4-87fb-4629-8b74-115e823503bb", "address": "fa:16:3e:57:29:40", "network": {"id": "a561a523-c9d8-4169-b449-5c0bdef7b53b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-751953866-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d716d19f79945db98f684df0cf4b302", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd3797d4-87", "ovs_interfaceid": "dd3797d4-87fb-4629-8b74-115e823503bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1059.425722] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:57:29:40', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '170f3b82-5915-4e36-bce9-4664ebb6be5e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dd3797d4-87fb-4629-8b74-115e823503bb', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1059.433566] env[62385]: DEBUG oslo.service.loopingcall [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1059.433814] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1059.434073] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-679b1c9d-7ef2-4270-ba0a-fbb38c2bc4bc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.455295] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1059.455295] env[62385]: value = "task-1206435" [ 1059.455295] env[62385]: _type = "Task" [ 1059.455295] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.465906] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206435, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.597486] env[62385]: DEBUG nova.compute.manager [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1059.626271] env[62385]: DEBUG nova.virt.hardware [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1059.626528] env[62385]: DEBUG nova.virt.hardware [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1059.626695] env[62385]: DEBUG nova.virt.hardware [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1059.626885] env[62385]: DEBUG nova.virt.hardware [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1059.627081] env[62385]: DEBUG nova.virt.hardware [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1059.627302] env[62385]: DEBUG nova.virt.hardware [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1059.628011] env[62385]: DEBUG nova.virt.hardware [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1059.628011] env[62385]: DEBUG nova.virt.hardware [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1059.628159] env[62385]: DEBUG nova.virt.hardware [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1059.628268] env[62385]: DEBUG nova.virt.hardware [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1059.628640] env[62385]: DEBUG nova.virt.hardware [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1059.629470] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c549df0a-ba92-4f80-a420-ada282adbf41 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.637281] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26be2f3f-4f86-4049-b5b6-7ff01829262d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.711620] env[62385]: DEBUG nova.compute.manager [req-a0df54a3-e336-4ec1-b805-7ce8e7d7240d req-ff99d0d4-87ec-46c9-b175-01d6d8f5028d service nova] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Received event network-changed-dd3797d4-87fb-4629-8b74-115e823503bb {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1059.711812] env[62385]: DEBUG nova.compute.manager [req-a0df54a3-e336-4ec1-b805-7ce8e7d7240d req-ff99d0d4-87ec-46c9-b175-01d6d8f5028d service nova] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Refreshing instance network info cache due to event network-changed-dd3797d4-87fb-4629-8b74-115e823503bb. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1059.712046] env[62385]: DEBUG oslo_concurrency.lockutils [req-a0df54a3-e336-4ec1-b805-7ce8e7d7240d req-ff99d0d4-87ec-46c9-b175-01d6d8f5028d service nova] Acquiring lock "refresh_cache-f93d9f5d-f7b6-4988-af7c-877a9e4c2104" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.713906] env[62385]: DEBUG oslo_concurrency.lockutils [req-a0df54a3-e336-4ec1-b805-7ce8e7d7240d req-ff99d0d4-87ec-46c9-b175-01d6d8f5028d service nova] Acquired lock "refresh_cache-f93d9f5d-f7b6-4988-af7c-877a9e4c2104" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.713906] env[62385]: DEBUG nova.network.neutron [req-a0df54a3-e336-4ec1-b805-7ce8e7d7240d req-ff99d0d4-87ec-46c9-b175-01d6d8f5028d service nova] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Refreshing network info cache for port dd3797d4-87fb-4629-8b74-115e823503bb {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1059.758255] env[62385]: DEBUG nova.network.neutron [req-1aeba257-8c93-4094-aa5c-15b2c4084fdb req-2e82a79e-6d07-4667-8c4f-f0c70f30c4f1 service nova] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Updated VIF entry in instance network info cache for port c308701b-9b9e-4ff3-9a01-c27aabf54f0d. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1059.758630] env[62385]: DEBUG nova.network.neutron [req-1aeba257-8c93-4094-aa5c-15b2c4084fdb req-2e82a79e-6d07-4667-8c4f-f0c70f30c4f1 service nova] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Updating instance_info_cache with network_info: [{"id": "c308701b-9b9e-4ff3-9a01-c27aabf54f0d", "address": "fa:16:3e:9a:ed:8c", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.210", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc308701b-9b", "ovs_interfaceid": "c308701b-9b9e-4ff3-9a01-c27aabf54f0d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1059.859449] env[62385]: DEBUG nova.compute.manager [req-6f51c44b-2677-47c3-a077-07289225ca90 req-5372abcd-a427-48fd-a300-9d44f84b92ae service nova] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Received event network-vif-plugged-331b6188-b103-437e-8436-5259bc356e31 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1059.859882] env[62385]: DEBUG oslo_concurrency.lockutils [req-6f51c44b-2677-47c3-a077-07289225ca90 req-5372abcd-a427-48fd-a300-9d44f84b92ae service nova] Acquiring lock "68748d25-017f-4b58-9984-70264abbc5ba-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.860138] env[62385]: DEBUG oslo_concurrency.lockutils [req-6f51c44b-2677-47c3-a077-07289225ca90 req-5372abcd-a427-48fd-a300-9d44f84b92ae service nova] Lock "68748d25-017f-4b58-9984-70264abbc5ba-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.860319] env[62385]: DEBUG oslo_concurrency.lockutils [req-6f51c44b-2677-47c3-a077-07289225ca90 req-5372abcd-a427-48fd-a300-9d44f84b92ae service nova] Lock "68748d25-017f-4b58-9984-70264abbc5ba-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.860726] env[62385]: DEBUG nova.compute.manager [req-6f51c44b-2677-47c3-a077-07289225ca90 req-5372abcd-a427-48fd-a300-9d44f84b92ae service nova] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] No waiting events found dispatching network-vif-plugged-331b6188-b103-437e-8436-5259bc356e31 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1059.860932] env[62385]: WARNING nova.compute.manager [req-6f51c44b-2677-47c3-a077-07289225ca90 req-5372abcd-a427-48fd-a300-9d44f84b92ae service nova] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Received unexpected event network-vif-plugged-331b6188-b103-437e-8436-5259bc356e31 for instance with vm_state active and task_state None. [ 1059.898453] env[62385]: DEBUG nova.network.neutron [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Successfully updated port: bc49aae7-4b0a-4cef-9b58-1baefe8a64ce {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1059.949362] env[62385]: DEBUG nova.network.neutron [None req-8bde1141-afeb-4c33-a510-88e0cc945e77 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Successfully updated port: 331b6188-b103-437e-8436-5259bc356e31 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1059.966772] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206435, 'name': CreateVM_Task, 'duration_secs': 0.342713} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.966935] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1059.967572] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.967742] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.968055] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1059.968300] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf12bc4b-5aa3-4b3f-9def-2d708be6f249 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.972625] env[62385]: DEBUG oslo_vmware.api [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for the task: (returnval){ [ 1059.972625] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]526c07bb-cc05-ea20-5620-cdc1f8501071" [ 1059.972625] env[62385]: _type = "Task" [ 1059.972625] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.980148] env[62385]: DEBUG oslo_vmware.api [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]526c07bb-cc05-ea20-5620-cdc1f8501071, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.261740] env[62385]: DEBUG oslo_concurrency.lockutils [req-1aeba257-8c93-4094-aa5c-15b2c4084fdb req-2e82a79e-6d07-4667-8c4f-f0c70f30c4f1 service nova] Releasing lock "refresh_cache-68748d25-017f-4b58-9984-70264abbc5ba" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.262026] env[62385]: DEBUG nova.compute.manager [req-1aeba257-8c93-4094-aa5c-15b2c4084fdb req-2e82a79e-6d07-4667-8c4f-f0c70f30c4f1 service nova] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Received event network-changed-b5615be2-bc51-4fe3-8905-4b7083550949 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1060.262301] env[62385]: DEBUG nova.compute.manager [req-1aeba257-8c93-4094-aa5c-15b2c4084fdb req-2e82a79e-6d07-4667-8c4f-f0c70f30c4f1 service nova] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Refreshing instance network info cache due to event network-changed-b5615be2-bc51-4fe3-8905-4b7083550949. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1060.262423] env[62385]: DEBUG oslo_concurrency.lockutils [req-1aeba257-8c93-4094-aa5c-15b2c4084fdb req-2e82a79e-6d07-4667-8c4f-f0c70f30c4f1 service nova] Acquiring lock "refresh_cache-5c9284ba-9f83-454f-b91b-9b0ea9ee3195" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.262570] env[62385]: DEBUG oslo_concurrency.lockutils [req-1aeba257-8c93-4094-aa5c-15b2c4084fdb req-2e82a79e-6d07-4667-8c4f-f0c70f30c4f1 service nova] Acquired lock "refresh_cache-5c9284ba-9f83-454f-b91b-9b0ea9ee3195" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.262735] env[62385]: DEBUG nova.network.neutron [req-1aeba257-8c93-4094-aa5c-15b2c4084fdb req-2e82a79e-6d07-4667-8c4f-f0c70f30c4f1 service nova] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Refreshing network info cache for port b5615be2-bc51-4fe3-8905-4b7083550949 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1060.402913] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Acquiring lock "refresh_cache-2cb0fee4-c922-402a-9251-10dc0da8c87d" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.402913] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Acquired lock "refresh_cache-2cb0fee4-c922-402a-9251-10dc0da8c87d" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.402913] env[62385]: DEBUG nova.network.neutron [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1060.451503] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8bde1141-afeb-4c33-a510-88e0cc945e77 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "refresh_cache-68748d25-017f-4b58-9984-70264abbc5ba" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.451723] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8bde1141-afeb-4c33-a510-88e0cc945e77 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquired lock "refresh_cache-68748d25-017f-4b58-9984-70264abbc5ba" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.452244] env[62385]: DEBUG nova.network.neutron [None req-8bde1141-afeb-4c33-a510-88e0cc945e77 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1060.460484] env[62385]: DEBUG nova.network.neutron [req-a0df54a3-e336-4ec1-b805-7ce8e7d7240d req-ff99d0d4-87ec-46c9-b175-01d6d8f5028d service nova] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Updated VIF entry in instance network info cache for port dd3797d4-87fb-4629-8b74-115e823503bb. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1060.460903] env[62385]: DEBUG nova.network.neutron [req-a0df54a3-e336-4ec1-b805-7ce8e7d7240d req-ff99d0d4-87ec-46c9-b175-01d6d8f5028d service nova] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Updating instance_info_cache with network_info: [{"id": "dd3797d4-87fb-4629-8b74-115e823503bb", "address": "fa:16:3e:57:29:40", "network": {"id": "a561a523-c9d8-4169-b449-5c0bdef7b53b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-751953866-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d716d19f79945db98f684df0cf4b302", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd3797d4-87", "ovs_interfaceid": "dd3797d4-87fb-4629-8b74-115e823503bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.482652] env[62385]: DEBUG oslo_vmware.api [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]526c07bb-cc05-ea20-5620-cdc1f8501071, 'name': SearchDatastore_Task, 'duration_secs': 0.010016} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.483032] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.483198] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1060.483432] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.483584] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.483765] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1060.484036] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-59a979cc-5262-4836-8370-5083b6ab0be6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.492467] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1060.492646] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1060.493381] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e5f96d0-64ed-4079-ba87-fa5875c2e95f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.498404] env[62385]: DEBUG oslo_vmware.api [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for the task: (returnval){ [ 1060.498404] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52df435f-6d1a-5ea2-2ffc-d9f83eb71dd6" [ 1060.498404] env[62385]: _type = "Task" [ 1060.498404] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.506505] env[62385]: DEBUG oslo_vmware.api [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52df435f-6d1a-5ea2-2ffc-d9f83eb71dd6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.935030] env[62385]: DEBUG nova.network.neutron [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1060.963578] env[62385]: DEBUG oslo_concurrency.lockutils [req-a0df54a3-e336-4ec1-b805-7ce8e7d7240d req-ff99d0d4-87ec-46c9-b175-01d6d8f5028d service nova] Releasing lock "refresh_cache-f93d9f5d-f7b6-4988-af7c-877a9e4c2104" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.003927] env[62385]: WARNING nova.network.neutron [None req-8bde1141-afeb-4c33-a510-88e0cc945e77 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] daad7070-dc06-4580-9f4a-ac8aaa546955 already exists in list: networks containing: ['daad7070-dc06-4580-9f4a-ac8aaa546955']. ignoring it [ 1061.006326] env[62385]: DEBUG nova.network.neutron [req-1aeba257-8c93-4094-aa5c-15b2c4084fdb req-2e82a79e-6d07-4667-8c4f-f0c70f30c4f1 service nova] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Updated VIF entry in instance network info cache for port b5615be2-bc51-4fe3-8905-4b7083550949. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1061.006660] env[62385]: DEBUG nova.network.neutron [req-1aeba257-8c93-4094-aa5c-15b2c4084fdb req-2e82a79e-6d07-4667-8c4f-f0c70f30c4f1 service nova] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Updating instance_info_cache with network_info: [{"id": "b5615be2-bc51-4fe3-8905-4b7083550949", "address": "fa:16:3e:dd:76:b3", "network": {"id": "afae6b07-d08c-413f-aa73-b545db1bb766", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1654965885-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "93111f5ba17c47de9691f90b99e03aa8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5615be2-bc", "ovs_interfaceid": "b5615be2-bc51-4fe3-8905-4b7083550949", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.014126] env[62385]: DEBUG oslo_vmware.api [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52df435f-6d1a-5ea2-2ffc-d9f83eb71dd6, 'name': SearchDatastore_Task, 'duration_secs': 0.008131} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.014880] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31b68070-e76a-4306-82ab-d3ae6ffca7f2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.020326] env[62385]: DEBUG oslo_vmware.api [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for the task: (returnval){ [ 1061.020326] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]524581f2-9252-6ee3-c6e2-a3abd549eb41" [ 1061.020326] env[62385]: _type = "Task" [ 1061.020326] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.027683] env[62385]: DEBUG oslo_vmware.api [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]524581f2-9252-6ee3-c6e2-a3abd549eb41, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.061108] env[62385]: DEBUG nova.network.neutron [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Updating instance_info_cache with network_info: [{"id": "bc49aae7-4b0a-4cef-9b58-1baefe8a64ce", "address": "fa:16:3e:82:f2:a1", "network": {"id": "d91cb364-52c9-4090-900e-3bdd2a57dc8b", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-443220934-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbdb8eed888d42129200077fc2ffb17f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69e41c97-4d75-4041-ae71-321e7e9d480b", "external-id": "nsx-vlan-transportzone-483", "segmentation_id": 483, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc49aae7-4b", "ovs_interfaceid": "bc49aae7-4b0a-4cef-9b58-1baefe8a64ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.350867] env[62385]: DEBUG nova.network.neutron [None req-8bde1141-afeb-4c33-a510-88e0cc945e77 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Updating instance_info_cache with network_info: [{"id": "c308701b-9b9e-4ff3-9a01-c27aabf54f0d", "address": "fa:16:3e:9a:ed:8c", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.210", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc308701b-9b", "ovs_interfaceid": "c308701b-9b9e-4ff3-9a01-c27aabf54f0d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "331b6188-b103-437e-8436-5259bc356e31", "address": "fa:16:3e:f4:84:82", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap331b6188-b1", "ovs_interfaceid": "331b6188-b103-437e-8436-5259bc356e31", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.509784] env[62385]: DEBUG oslo_concurrency.lockutils [req-1aeba257-8c93-4094-aa5c-15b2c4084fdb req-2e82a79e-6d07-4667-8c4f-f0c70f30c4f1 service nova] Releasing lock "refresh_cache-5c9284ba-9f83-454f-b91b-9b0ea9ee3195" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.530566] env[62385]: DEBUG oslo_vmware.api [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]524581f2-9252-6ee3-c6e2-a3abd549eb41, 'name': SearchDatastore_Task, 'duration_secs': 0.009022} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.530872] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.531167] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] f93d9f5d-f7b6-4988-af7c-877a9e4c2104/f93d9f5d-f7b6-4988-af7c-877a9e4c2104.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1061.531435] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-05d1125f-1073-4189-b351-29c06fcafdd6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.538941] env[62385]: DEBUG oslo_vmware.api [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for the task: (returnval){ [ 1061.538941] env[62385]: value = "task-1206437" [ 1061.538941] env[62385]: _type = "Task" [ 1061.538941] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.546235] env[62385]: DEBUG oslo_vmware.api [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206437, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.563765] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Releasing lock "refresh_cache-2cb0fee4-c922-402a-9251-10dc0da8c87d" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.564103] env[62385]: DEBUG nova.compute.manager [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Instance network_info: |[{"id": "bc49aae7-4b0a-4cef-9b58-1baefe8a64ce", "address": "fa:16:3e:82:f2:a1", "network": {"id": "d91cb364-52c9-4090-900e-3bdd2a57dc8b", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-443220934-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbdb8eed888d42129200077fc2ffb17f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69e41c97-4d75-4041-ae71-321e7e9d480b", "external-id": "nsx-vlan-transportzone-483", "segmentation_id": 483, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc49aae7-4b", "ovs_interfaceid": "bc49aae7-4b0a-4cef-9b58-1baefe8a64ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1061.564486] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:82:f2:a1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69e41c97-4d75-4041-ae71-321e7e9d480b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bc49aae7-4b0a-4cef-9b58-1baefe8a64ce', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1061.571900] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Creating folder: Project (cbdb8eed888d42129200077fc2ffb17f). Parent ref: group-v261107. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1061.572169] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2efc73c1-4094-4bce-8771-81bb6f78db54 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.583744] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Created folder: Project (cbdb8eed888d42129200077fc2ffb17f) in parent group-v261107. [ 1061.583940] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Creating folder: Instances. Parent ref: group-v261308. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1061.584189] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4ac33d0a-ac6a-4571-9b9a-0da879fd11eb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.595417] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Created folder: Instances in parent group-v261308. [ 1061.595657] env[62385]: DEBUG oslo.service.loopingcall [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1061.595847] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1061.596154] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-457c82fe-1471-4ee4-9cd8-d6344e4c6fa0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.614118] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1061.614118] env[62385]: value = "task-1206440" [ 1061.614118] env[62385]: _type = "Task" [ 1061.614118] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.621512] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206440, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.740122] env[62385]: DEBUG nova.compute.manager [req-15478577-8b4b-4096-86c2-921e32bac3be req-cb042dec-4fcc-44b3-a115-386403911bb3 service nova] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Received event network-vif-plugged-bc49aae7-4b0a-4cef-9b58-1baefe8a64ce {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1061.740122] env[62385]: DEBUG oslo_concurrency.lockutils [req-15478577-8b4b-4096-86c2-921e32bac3be req-cb042dec-4fcc-44b3-a115-386403911bb3 service nova] Acquiring lock "2cb0fee4-c922-402a-9251-10dc0da8c87d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.740212] env[62385]: DEBUG oslo_concurrency.lockutils [req-15478577-8b4b-4096-86c2-921e32bac3be req-cb042dec-4fcc-44b3-a115-386403911bb3 service nova] Lock "2cb0fee4-c922-402a-9251-10dc0da8c87d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.740320] env[62385]: DEBUG oslo_concurrency.lockutils [req-15478577-8b4b-4096-86c2-921e32bac3be req-cb042dec-4fcc-44b3-a115-386403911bb3 service nova] Lock "2cb0fee4-c922-402a-9251-10dc0da8c87d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.740490] env[62385]: DEBUG nova.compute.manager [req-15478577-8b4b-4096-86c2-921e32bac3be req-cb042dec-4fcc-44b3-a115-386403911bb3 service nova] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] No waiting events found dispatching network-vif-plugged-bc49aae7-4b0a-4cef-9b58-1baefe8a64ce {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1061.740677] env[62385]: WARNING nova.compute.manager [req-15478577-8b4b-4096-86c2-921e32bac3be req-cb042dec-4fcc-44b3-a115-386403911bb3 service nova] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Received unexpected event network-vif-plugged-bc49aae7-4b0a-4cef-9b58-1baefe8a64ce for instance with vm_state building and task_state spawning. [ 1061.740847] env[62385]: DEBUG nova.compute.manager [req-15478577-8b4b-4096-86c2-921e32bac3be req-cb042dec-4fcc-44b3-a115-386403911bb3 service nova] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Received event network-changed-bc49aae7-4b0a-4cef-9b58-1baefe8a64ce {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1061.741014] env[62385]: DEBUG nova.compute.manager [req-15478577-8b4b-4096-86c2-921e32bac3be req-cb042dec-4fcc-44b3-a115-386403911bb3 service nova] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Refreshing instance network info cache due to event network-changed-bc49aae7-4b0a-4cef-9b58-1baefe8a64ce. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1061.741212] env[62385]: DEBUG oslo_concurrency.lockutils [req-15478577-8b4b-4096-86c2-921e32bac3be req-cb042dec-4fcc-44b3-a115-386403911bb3 service nova] Acquiring lock "refresh_cache-2cb0fee4-c922-402a-9251-10dc0da8c87d" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1061.741347] env[62385]: DEBUG oslo_concurrency.lockutils [req-15478577-8b4b-4096-86c2-921e32bac3be req-cb042dec-4fcc-44b3-a115-386403911bb3 service nova] Acquired lock "refresh_cache-2cb0fee4-c922-402a-9251-10dc0da8c87d" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.741504] env[62385]: DEBUG nova.network.neutron [req-15478577-8b4b-4096-86c2-921e32bac3be req-cb042dec-4fcc-44b3-a115-386403911bb3 service nova] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Refreshing network info cache for port bc49aae7-4b0a-4cef-9b58-1baefe8a64ce {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1061.853517] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8bde1141-afeb-4c33-a510-88e0cc945e77 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Releasing lock "refresh_cache-68748d25-017f-4b58-9984-70264abbc5ba" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.854192] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8bde1141-afeb-4c33-a510-88e0cc945e77 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "68748d25-017f-4b58-9984-70264abbc5ba" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1061.854364] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8bde1141-afeb-4c33-a510-88e0cc945e77 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquired lock "68748d25-017f-4b58-9984-70264abbc5ba" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.855232] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f81db5a-18f2-466a-abd2-75d342e50a4d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.873025] env[62385]: DEBUG nova.virt.hardware [None req-8bde1141-afeb-4c33-a510-88e0cc945e77 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1061.873306] env[62385]: DEBUG nova.virt.hardware [None req-8bde1141-afeb-4c33-a510-88e0cc945e77 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1061.873483] env[62385]: DEBUG nova.virt.hardware [None req-8bde1141-afeb-4c33-a510-88e0cc945e77 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1061.873681] env[62385]: DEBUG nova.virt.hardware [None req-8bde1141-afeb-4c33-a510-88e0cc945e77 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1061.873827] env[62385]: DEBUG nova.virt.hardware [None req-8bde1141-afeb-4c33-a510-88e0cc945e77 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1061.873981] env[62385]: DEBUG nova.virt.hardware [None req-8bde1141-afeb-4c33-a510-88e0cc945e77 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1061.874217] env[62385]: DEBUG nova.virt.hardware [None req-8bde1141-afeb-4c33-a510-88e0cc945e77 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1061.874392] env[62385]: DEBUG nova.virt.hardware [None req-8bde1141-afeb-4c33-a510-88e0cc945e77 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1061.874583] env[62385]: DEBUG nova.virt.hardware [None req-8bde1141-afeb-4c33-a510-88e0cc945e77 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1061.874706] env[62385]: DEBUG nova.virt.hardware [None req-8bde1141-afeb-4c33-a510-88e0cc945e77 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1061.874884] env[62385]: DEBUG nova.virt.hardware [None req-8bde1141-afeb-4c33-a510-88e0cc945e77 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1061.881196] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8bde1141-afeb-4c33-a510-88e0cc945e77 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Reconfiguring VM to attach interface {{(pid=62385) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 1061.882090] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-29274bc8-f230-4ed1-94f0-acf9a108eb3d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.895999] env[62385]: DEBUG nova.compute.manager [req-89181130-89c1-438e-8d9d-9ce91f9aa0ee req-d31a5d32-cf9b-4799-85b5-0185e51d6499 service nova] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Received event network-changed-331b6188-b103-437e-8436-5259bc356e31 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1061.896196] env[62385]: DEBUG nova.compute.manager [req-89181130-89c1-438e-8d9d-9ce91f9aa0ee req-d31a5d32-cf9b-4799-85b5-0185e51d6499 service nova] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Refreshing instance network info cache due to event network-changed-331b6188-b103-437e-8436-5259bc356e31. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1061.896429] env[62385]: DEBUG oslo_concurrency.lockutils [req-89181130-89c1-438e-8d9d-9ce91f9aa0ee req-d31a5d32-cf9b-4799-85b5-0185e51d6499 service nova] Acquiring lock "refresh_cache-68748d25-017f-4b58-9984-70264abbc5ba" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1061.896562] env[62385]: DEBUG oslo_concurrency.lockutils [req-89181130-89c1-438e-8d9d-9ce91f9aa0ee req-d31a5d32-cf9b-4799-85b5-0185e51d6499 service nova] Acquired lock "refresh_cache-68748d25-017f-4b58-9984-70264abbc5ba" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.896724] env[62385]: DEBUG nova.network.neutron [req-89181130-89c1-438e-8d9d-9ce91f9aa0ee req-d31a5d32-cf9b-4799-85b5-0185e51d6499 service nova] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Refreshing network info cache for port 331b6188-b103-437e-8436-5259bc356e31 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1061.903508] env[62385]: DEBUG oslo_vmware.api [None req-8bde1141-afeb-4c33-a510-88e0cc945e77 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 1061.903508] env[62385]: value = "task-1206441" [ 1061.903508] env[62385]: _type = "Task" [ 1061.903508] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.912060] env[62385]: DEBUG oslo_vmware.api [None req-8bde1141-afeb-4c33-a510-88e0cc945e77 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206441, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.048024] env[62385]: DEBUG oslo_vmware.api [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206437, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.124815] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206440, 'name': CreateVM_Task, 'duration_secs': 0.390476} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.124995] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1062.125720] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1062.125896] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.126268] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1062.126528] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91ca6711-38ae-4026-ac79-6f41185891f9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.130863] env[62385]: DEBUG oslo_vmware.api [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Waiting for the task: (returnval){ [ 1062.130863] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5257638f-cbdd-b5d9-6b47-4310d346396a" [ 1062.130863] env[62385]: _type = "Task" [ 1062.130863] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.138152] env[62385]: DEBUG oslo_vmware.api [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5257638f-cbdd-b5d9-6b47-4310d346396a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.414119] env[62385]: DEBUG oslo_vmware.api [None req-8bde1141-afeb-4c33-a510-88e0cc945e77 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206441, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.509032] env[62385]: DEBUG nova.network.neutron [req-15478577-8b4b-4096-86c2-921e32bac3be req-cb042dec-4fcc-44b3-a115-386403911bb3 service nova] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Updated VIF entry in instance network info cache for port bc49aae7-4b0a-4cef-9b58-1baefe8a64ce. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1062.509032] env[62385]: DEBUG nova.network.neutron [req-15478577-8b4b-4096-86c2-921e32bac3be req-cb042dec-4fcc-44b3-a115-386403911bb3 service nova] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Updating instance_info_cache with network_info: [{"id": "bc49aae7-4b0a-4cef-9b58-1baefe8a64ce", "address": "fa:16:3e:82:f2:a1", "network": {"id": "d91cb364-52c9-4090-900e-3bdd2a57dc8b", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-443220934-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbdb8eed888d42129200077fc2ffb17f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69e41c97-4d75-4041-ae71-321e7e9d480b", "external-id": "nsx-vlan-transportzone-483", "segmentation_id": 483, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc49aae7-4b", "ovs_interfaceid": "bc49aae7-4b0a-4cef-9b58-1baefe8a64ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1062.556782] env[62385]: DEBUG oslo_vmware.api [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206437, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.641996] env[62385]: DEBUG oslo_vmware.api [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5257638f-cbdd-b5d9-6b47-4310d346396a, 'name': SearchDatastore_Task, 'duration_secs': 0.059096} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.642365] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.642582] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1062.642817] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1062.642968] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.643165] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1062.643458] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c65872e5-f1ea-4f62-870e-aa853b721cba {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.650696] env[62385]: DEBUG nova.network.neutron [req-89181130-89c1-438e-8d9d-9ce91f9aa0ee req-d31a5d32-cf9b-4799-85b5-0185e51d6499 service nova] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Updated VIF entry in instance network info cache for port 331b6188-b103-437e-8436-5259bc356e31. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1062.651285] env[62385]: DEBUG nova.network.neutron [req-89181130-89c1-438e-8d9d-9ce91f9aa0ee req-d31a5d32-cf9b-4799-85b5-0185e51d6499 service nova] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Updating instance_info_cache with network_info: [{"id": "c308701b-9b9e-4ff3-9a01-c27aabf54f0d", "address": "fa:16:3e:9a:ed:8c", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.210", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc308701b-9b", "ovs_interfaceid": "c308701b-9b9e-4ff3-9a01-c27aabf54f0d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "331b6188-b103-437e-8436-5259bc356e31", "address": "fa:16:3e:f4:84:82", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap331b6188-b1", "ovs_interfaceid": "331b6188-b103-437e-8436-5259bc356e31", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1062.656994] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1062.657181] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1062.658296] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb7c14ce-562d-410c-9d4b-cb7d16c398c6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.665287] env[62385]: DEBUG oslo_vmware.api [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Waiting for the task: (returnval){ [ 1062.665287] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52690091-f6d5-24be-5569-85979cf14d91" [ 1062.665287] env[62385]: _type = "Task" [ 1062.665287] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.674269] env[62385]: DEBUG oslo_vmware.api [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52690091-f6d5-24be-5569-85979cf14d91, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.914679] env[62385]: DEBUG oslo_vmware.api [None req-8bde1141-afeb-4c33-a510-88e0cc945e77 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206441, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.014752] env[62385]: DEBUG oslo_concurrency.lockutils [req-15478577-8b4b-4096-86c2-921e32bac3be req-cb042dec-4fcc-44b3-a115-386403911bb3 service nova] Releasing lock "refresh_cache-2cb0fee4-c922-402a-9251-10dc0da8c87d" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.058151] env[62385]: DEBUG oslo_vmware.api [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206437, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.311231} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.058676] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] f93d9f5d-f7b6-4988-af7c-877a9e4c2104/f93d9f5d-f7b6-4988-af7c-877a9e4c2104.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1063.058676] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1063.058925] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b506399c-100c-459e-9b91-4af3e78ad60f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.070358] env[62385]: DEBUG oslo_vmware.api [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for the task: (returnval){ [ 1063.070358] env[62385]: value = "task-1206442" [ 1063.070358] env[62385]: _type = "Task" [ 1063.070358] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.082232] env[62385]: DEBUG oslo_vmware.api [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206442, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.154193] env[62385]: DEBUG oslo_concurrency.lockutils [req-89181130-89c1-438e-8d9d-9ce91f9aa0ee req-d31a5d32-cf9b-4799-85b5-0185e51d6499 service nova] Releasing lock "refresh_cache-68748d25-017f-4b58-9984-70264abbc5ba" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.179212] env[62385]: DEBUG oslo_vmware.api [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52690091-f6d5-24be-5569-85979cf14d91, 'name': SearchDatastore_Task, 'duration_secs': 0.052095} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.180095] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4cd12dbd-3a33-40bc-bb4a-185fead6d3ee {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.185458] env[62385]: DEBUG oslo_vmware.api [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Waiting for the task: (returnval){ [ 1063.185458] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e40caa-d5d9-011d-34c9-c6772e468c9d" [ 1063.185458] env[62385]: _type = "Task" [ 1063.185458] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.192900] env[62385]: DEBUG oslo_vmware.api [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e40caa-d5d9-011d-34c9-c6772e468c9d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.415356] env[62385]: DEBUG oslo_vmware.api [None req-8bde1141-afeb-4c33-a510-88e0cc945e77 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206441, 'name': ReconfigVM_Task, 'duration_secs': 1.272494} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.416665] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8bde1141-afeb-4c33-a510-88e0cc945e77 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Releasing lock "68748d25-017f-4b58-9984-70264abbc5ba" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.416665] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8bde1141-afeb-4c33-a510-88e0cc945e77 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Reconfigured VM to attach interface {{(pid=62385) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 1063.579502] env[62385]: DEBUG oslo_vmware.api [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206442, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.087481} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.579829] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1063.581392] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf9092a2-75e6-4605-bb50-53c791ac07b3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.603776] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] f93d9f5d-f7b6-4988-af7c-877a9e4c2104/f93d9f5d-f7b6-4988-af7c-877a9e4c2104.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1063.604199] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-79ef3aec-a0dd-46cf-8b78-b5527f861c60 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.623741] env[62385]: DEBUG oslo_vmware.api [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for the task: (returnval){ [ 1063.623741] env[62385]: value = "task-1206444" [ 1063.623741] env[62385]: _type = "Task" [ 1063.623741] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.631450] env[62385]: DEBUG oslo_vmware.api [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206444, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.700287] env[62385]: DEBUG oslo_vmware.api [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e40caa-d5d9-011d-34c9-c6772e468c9d, 'name': SearchDatastore_Task, 'duration_secs': 0.015686} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.702120] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.702120] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 2cb0fee4-c922-402a-9251-10dc0da8c87d/2cb0fee4-c922-402a-9251-10dc0da8c87d.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1063.702120] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1ddee7f6-c618-4396-b6cf-d34d7cd5808d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.716148] env[62385]: DEBUG oslo_vmware.api [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Waiting for the task: (returnval){ [ 1063.716148] env[62385]: value = "task-1206445" [ 1063.716148] env[62385]: _type = "Task" [ 1063.716148] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.728137] env[62385]: DEBUG oslo_vmware.api [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Task: {'id': task-1206445, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.922129] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8bde1141-afeb-4c33-a510-88e0cc945e77 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "interface-68748d25-017f-4b58-9984-70264abbc5ba-331b6188-b103-437e-8436-5259bc356e31" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.740s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.030808] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1064.031722] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1064.031722] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Starting heal instance info cache {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1064.031722] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Rebuilding the list of instances to heal {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1064.133777] env[62385]: DEBUG oslo_vmware.api [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206444, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.226636] env[62385]: DEBUG oslo_vmware.api [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Task: {'id': task-1206445, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.537174] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Skipping network cache update for instance because it is Building. {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 1064.537317] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Skipping network cache update for instance because it is Building. {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 1064.574612] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquiring lock "refresh_cache-9d6f098a-0b05-43ef-96b0-9eb99ad3538c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.574800] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquired lock "refresh_cache-9d6f098a-0b05-43ef-96b0-9eb99ad3538c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.574954] env[62385]: DEBUG nova.network.neutron [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Forcefully refreshing network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1064.575127] env[62385]: DEBUG nova.objects.instance [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lazy-loading 'info_cache' on Instance uuid 9d6f098a-0b05-43ef-96b0-9eb99ad3538c {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1064.634627] env[62385]: DEBUG oslo_vmware.api [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206444, 'name': ReconfigVM_Task, 'duration_secs': 0.657444} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.634922] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Reconfigured VM instance instance-0000006e to attach disk [datastore2] f93d9f5d-f7b6-4988-af7c-877a9e4c2104/f93d9f5d-f7b6-4988-af7c-877a9e4c2104.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1064.635574] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c73f23a2-e652-4c7e-817d-0917224df514 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.641927] env[62385]: DEBUG oslo_vmware.api [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for the task: (returnval){ [ 1064.641927] env[62385]: value = "task-1206446" [ 1064.641927] env[62385]: _type = "Task" [ 1064.641927] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.649705] env[62385]: DEBUG oslo_vmware.api [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206446, 'name': Rename_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.727272] env[62385]: DEBUG oslo_vmware.api [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Task: {'id': task-1206445, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.518285} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.727583] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 2cb0fee4-c922-402a-9251-10dc0da8c87d/2cb0fee4-c922-402a-9251-10dc0da8c87d.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1064.727883] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1064.728199] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-54b0f2f8-bd2e-49ae-a849-c7374d1a5e32 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.735368] env[62385]: DEBUG oslo_vmware.api [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Waiting for the task: (returnval){ [ 1064.735368] env[62385]: value = "task-1206447" [ 1064.735368] env[62385]: _type = "Task" [ 1064.735368] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.744630] env[62385]: DEBUG oslo_vmware.api [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Task: {'id': task-1206447, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.151575] env[62385]: DEBUG oslo_vmware.api [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206446, 'name': Rename_Task, 'duration_secs': 0.300261} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.152029] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1065.152569] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b8969b03-d436-4ae0-a1a9-0844701c1e83 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.159128] env[62385]: DEBUG oslo_vmware.api [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for the task: (returnval){ [ 1065.159128] env[62385]: value = "task-1206448" [ 1065.159128] env[62385]: _type = "Task" [ 1065.159128] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.166477] env[62385]: DEBUG oslo_vmware.api [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206448, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.244455] env[62385]: DEBUG oslo_vmware.api [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Task: {'id': task-1206447, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069112} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.244773] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1065.245549] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a40e2ce9-d956-4215-8f43-8ad47dc30b20 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.270051] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] 2cb0fee4-c922-402a-9251-10dc0da8c87d/2cb0fee4-c922-402a-9251-10dc0da8c87d.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1065.270315] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f85ad53d-dc76-465a-b3fd-81b1f9fa493f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.290104] env[62385]: DEBUG oslo_vmware.api [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Waiting for the task: (returnval){ [ 1065.290104] env[62385]: value = "task-1206449" [ 1065.290104] env[62385]: _type = "Task" [ 1065.290104] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.298094] env[62385]: DEBUG oslo_vmware.api [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Task: {'id': task-1206449, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.391498] env[62385]: DEBUG oslo_concurrency.lockutils [None req-af1917ad-a569-4b83-b638-ab28fd00865a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "interface-68748d25-017f-4b58-9984-70264abbc5ba-331b6188-b103-437e-8436-5259bc356e31" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.391812] env[62385]: DEBUG oslo_concurrency.lockutils [None req-af1917ad-a569-4b83-b638-ab28fd00865a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "interface-68748d25-017f-4b58-9984-70264abbc5ba-331b6188-b103-437e-8436-5259bc356e31" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.670086] env[62385]: DEBUG oslo_vmware.api [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206448, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.800300] env[62385]: DEBUG oslo_vmware.api [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Task: {'id': task-1206449, 'name': ReconfigVM_Task, 'duration_secs': 0.353116} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.800617] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Reconfigured VM instance instance-0000006f to attach disk [datastore2] 2cb0fee4-c922-402a-9251-10dc0da8c87d/2cb0fee4-c922-402a-9251-10dc0da8c87d.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1065.801335] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c22010be-62de-4892-94cc-19af8dcb9018 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.808042] env[62385]: DEBUG oslo_vmware.api [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Waiting for the task: (returnval){ [ 1065.808042] env[62385]: value = "task-1206451" [ 1065.808042] env[62385]: _type = "Task" [ 1065.808042] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.816019] env[62385]: DEBUG oslo_vmware.api [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Task: {'id': task-1206451, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.894398] env[62385]: DEBUG oslo_concurrency.lockutils [None req-af1917ad-a569-4b83-b638-ab28fd00865a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "68748d25-017f-4b58-9984-70264abbc5ba" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.894614] env[62385]: DEBUG oslo_concurrency.lockutils [None req-af1917ad-a569-4b83-b638-ab28fd00865a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquired lock "68748d25-017f-4b58-9984-70264abbc5ba" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.895533] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e7339a4-0017-4f36-a8fb-35613444da75 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.912742] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-713aa40d-3307-45b2-b34a-7de0b70be0e5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.938553] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-af1917ad-a569-4b83-b638-ab28fd00865a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Reconfiguring VM to detach interface {{(pid=62385) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 1065.938656] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-463020b3-33d1-42f9-b0ff-9b11a923abf7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.956171] env[62385]: DEBUG oslo_vmware.api [None req-af1917ad-a569-4b83-b638-ab28fd00865a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 1065.956171] env[62385]: value = "task-1206452" [ 1065.956171] env[62385]: _type = "Task" [ 1065.956171] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.965574] env[62385]: DEBUG oslo_vmware.api [None req-af1917ad-a569-4b83-b638-ab28fd00865a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206452, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.169749] env[62385]: DEBUG oslo_vmware.api [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206448, 'name': PowerOnVM_Task, 'duration_secs': 0.562653} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.170092] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1066.170136] env[62385]: INFO nova.compute.manager [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Took 8.84 seconds to spawn the instance on the hypervisor. [ 1066.171061] env[62385]: DEBUG nova.compute.manager [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1066.171178] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-789833dd-38fc-496c-9836-59a99f0cd7aa {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.294775] env[62385]: DEBUG nova.network.neutron [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Updating instance_info_cache with network_info: [{"id": "ff03b686-6596-418e-b748-0a8a6799a2b5", "address": "fa:16:3e:91:05:f2", "network": {"id": "afae6b07-d08c-413f-aa73-b545db1bb766", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1654965885-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.163", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "93111f5ba17c47de9691f90b99e03aa8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff03b686-65", "ovs_interfaceid": "ff03b686-6596-418e-b748-0a8a6799a2b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1066.318425] env[62385]: DEBUG oslo_vmware.api [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Task: {'id': task-1206451, 'name': Rename_Task, 'duration_secs': 0.142242} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.318713] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1066.318989] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5388a791-bb7e-4bb1-9767-9d716a4dfb0a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.325890] env[62385]: DEBUG oslo_vmware.api [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Waiting for the task: (returnval){ [ 1066.325890] env[62385]: value = "task-1206453" [ 1066.325890] env[62385]: _type = "Task" [ 1066.325890] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.334148] env[62385]: DEBUG oslo_vmware.api [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Task: {'id': task-1206453, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.466328] env[62385]: DEBUG oslo_vmware.api [None req-af1917ad-a569-4b83-b638-ab28fd00865a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206452, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.687411] env[62385]: INFO nova.compute.manager [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Took 14.18 seconds to build instance. [ 1066.797108] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Releasing lock "refresh_cache-9d6f098a-0b05-43ef-96b0-9eb99ad3538c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1066.797322] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Updated the network info_cache for instance {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1066.797513] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1066.797677] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1066.797823] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1066.797973] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1066.798131] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1066.798280] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1066.798409] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62385) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1066.798552] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager.update_available_resource {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1066.836243] env[62385]: DEBUG oslo_vmware.api [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Task: {'id': task-1206453, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.967159] env[62385]: DEBUG oslo_vmware.api [None req-af1917ad-a569-4b83-b638-ab28fd00865a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206452, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.189771] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c9464585-a1b8-499b-9e10-f61841291de5 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Lock "f93d9f5d-f7b6-4988-af7c-877a9e4c2104" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.081s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.301681] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.301925] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.302105] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.302263] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62385) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1067.303152] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8fa9fbb-79fc-47e4-8798-ca6e504ba908 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.312284] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65877d52-5d1c-4694-b25f-b86321a8aa22 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.330413] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c51555a-e8cd-492e-9222-1cc6c7228263 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.337178] env[62385]: DEBUG oslo_vmware.api [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Task: {'id': task-1206453, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.339559] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d3cb337-8d98-4f45-8710-5066383e90e8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.369647] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180784MB free_disk=155GB free_vcpus=48 pci_devices=None {{(pid=62385) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1067.369839] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.369991] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.468587] env[62385]: DEBUG oslo_vmware.api [None req-af1917ad-a569-4b83-b638-ab28fd00865a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206452, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.781520] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0a480391-987b-4664-8241-e43a8295ee9e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Acquiring lock "f93d9f5d-f7b6-4988-af7c-877a9e4c2104" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.781787] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0a480391-987b-4664-8241-e43a8295ee9e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Lock "f93d9f5d-f7b6-4988-af7c-877a9e4c2104" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.782065] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0a480391-987b-4664-8241-e43a8295ee9e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Acquiring lock "f93d9f5d-f7b6-4988-af7c-877a9e4c2104-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.782287] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0a480391-987b-4664-8241-e43a8295ee9e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Lock "f93d9f5d-f7b6-4988-af7c-877a9e4c2104-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.782480] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0a480391-987b-4664-8241-e43a8295ee9e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Lock "f93d9f5d-f7b6-4988-af7c-877a9e4c2104-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.784773] env[62385]: INFO nova.compute.manager [None req-0a480391-987b-4664-8241-e43a8295ee9e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Terminating instance [ 1067.786625] env[62385]: DEBUG nova.compute.manager [None req-0a480391-987b-4664-8241-e43a8295ee9e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1067.786837] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0a480391-987b-4664-8241-e43a8295ee9e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1067.787902] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ee78b13-b502-4f7e-935a-f7abc6ca64a0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.795606] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a480391-987b-4664-8241-e43a8295ee9e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1067.795851] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fb015f4f-78ee-4bd3-9e34-f4f830730c24 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.801894] env[62385]: DEBUG oslo_vmware.api [None req-0a480391-987b-4664-8241-e43a8295ee9e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for the task: (returnval){ [ 1067.801894] env[62385]: value = "task-1206455" [ 1067.801894] env[62385]: _type = "Task" [ 1067.801894] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.812164] env[62385]: DEBUG oslo_vmware.api [None req-0a480391-987b-4664-8241-e43a8295ee9e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206455, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.836542] env[62385]: DEBUG oslo_vmware.api [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Task: {'id': task-1206453, 'name': PowerOnVM_Task, 'duration_secs': 1.420441} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.836826] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1067.837052] env[62385]: INFO nova.compute.manager [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Took 8.24 seconds to spawn the instance on the hypervisor. [ 1067.837245] env[62385]: DEBUG nova.compute.manager [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1067.838075] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6e01069-cb71-4655-93a9-f640c68ea409 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.969284] env[62385]: DEBUG oslo_vmware.api [None req-af1917ad-a569-4b83-b638-ab28fd00865a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206452, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.314359] env[62385]: DEBUG oslo_vmware.api [None req-0a480391-987b-4664-8241-e43a8295ee9e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206455, 'name': PowerOffVM_Task, 'duration_secs': 0.3291} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.314821] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a480391-987b-4664-8241-e43a8295ee9e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1068.314923] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0a480391-987b-4664-8241-e43a8295ee9e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1068.315245] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-32164904-269c-4ea1-992d-acef9b9326cd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.354917] env[62385]: INFO nova.compute.manager [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Took 14.36 seconds to build instance. [ 1068.383151] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0a480391-987b-4664-8241-e43a8295ee9e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1068.383443] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0a480391-987b-4664-8241-e43a8295ee9e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Deleting contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1068.383683] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a480391-987b-4664-8241-e43a8295ee9e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Deleting the datastore file [datastore2] f93d9f5d-f7b6-4988-af7c-877a9e4c2104 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1068.385110] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b6cf0412-8209-460a-9fdd-73802e964281 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.393294] env[62385]: DEBUG oslo_vmware.api [None req-0a480391-987b-4664-8241-e43a8295ee9e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for the task: (returnval){ [ 1068.393294] env[62385]: value = "task-1206457" [ 1068.393294] env[62385]: _type = "Task" [ 1068.393294] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.402842] env[62385]: DEBUG oslo_vmware.api [None req-0a480391-987b-4664-8241-e43a8295ee9e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206457, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.468611] env[62385]: DEBUG oslo_vmware.api [None req-af1917ad-a569-4b83-b638-ab28fd00865a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206452, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.480522] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 9d6f098a-0b05-43ef-96b0-9eb99ad3538c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1068.480723] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 57637d6e-39e2-405c-bbcf-64bf0ffbebd1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1068.480809] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 68748d25-017f-4b58-9984-70264abbc5ba actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1068.480932] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 8e1d416c-2624-468c-94d7-6265e9f4178f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1068.481064] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1068.481183] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance c49d4366-2ea0-44f2-8463-daf62bd1c40c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1068.481297] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1068.481406] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 5c9284ba-9f83-454f-b91b-9b0ea9ee3195 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1068.481515] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance f93d9f5d-f7b6-4988-af7c-877a9e4c2104 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1068.481621] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 2cb0fee4-c922-402a-9251-10dc0da8c87d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1068.481823] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Total usable vcpus: 48, total allocated vcpus: 10 {{(pid=62385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1068.481955] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2432MB phys_disk=200GB used_disk=9GB total_vcpus=48 used_vcpus=10 pci_stats=[] {{(pid=62385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1068.613624] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a70e880b-bce6-49b7-941c-0286b6bd8c9b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.621301] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-000d905f-0d3a-4229-a96b-98c83000d1da {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.652019] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-086c585e-365b-4fbb-a514-537d7909f23b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.659210] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9c5040c-44c8-4afa-907c-a35c479cde60 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.671782] env[62385]: DEBUG nova.compute.provider_tree [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1068.856982] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8278d82f-96de-4f70-bd53-f0fd2329dd78 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Lock "2cb0fee4-c922-402a-9251-10dc0da8c87d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.869s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.903265] env[62385]: DEBUG oslo_vmware.api [None req-0a480391-987b-4664-8241-e43a8295ee9e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206457, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.246843} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.903512] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a480391-987b-4664-8241-e43a8295ee9e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1068.903701] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0a480391-987b-4664-8241-e43a8295ee9e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Deleted contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1068.903879] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0a480391-987b-4664-8241-e43a8295ee9e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1068.904067] env[62385]: INFO nova.compute.manager [None req-0a480391-987b-4664-8241-e43a8295ee9e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1068.904308] env[62385]: DEBUG oslo.service.loopingcall [None req-0a480391-987b-4664-8241-e43a8295ee9e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1068.904500] env[62385]: DEBUG nova.compute.manager [-] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1068.904593] env[62385]: DEBUG nova.network.neutron [-] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1068.968026] env[62385]: DEBUG oslo_vmware.api [None req-af1917ad-a569-4b83-b638-ab28fd00865a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206452, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.146254] env[62385]: DEBUG nova.compute.manager [req-6d33e511-42c9-41ed-8e25-ff29daf9fe69 req-61599d73-67cc-4f24-b5d8-9b0b434383e8 service nova] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Received event network-vif-deleted-dd3797d4-87fb-4629-8b74-115e823503bb {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1069.146352] env[62385]: INFO nova.compute.manager [req-6d33e511-42c9-41ed-8e25-ff29daf9fe69 req-61599d73-67cc-4f24-b5d8-9b0b434383e8 service nova] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Neutron deleted interface dd3797d4-87fb-4629-8b74-115e823503bb; detaching it from the instance and deleting it from the info cache [ 1069.146515] env[62385]: DEBUG nova.network.neutron [req-6d33e511-42c9-41ed-8e25-ff29daf9fe69 req-61599d73-67cc-4f24-b5d8-9b0b434383e8 service nova] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1069.179360] env[62385]: DEBUG nova.scheduler.client.report [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1069.182902] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a7cb0fd8-040c-4daa-907f-4e312118d1f3 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Acquiring lock "2cb0fee4-c922-402a-9251-10dc0da8c87d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.182902] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a7cb0fd8-040c-4daa-907f-4e312118d1f3 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Lock "2cb0fee4-c922-402a-9251-10dc0da8c87d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.182902] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a7cb0fd8-040c-4daa-907f-4e312118d1f3 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Acquiring lock "2cb0fee4-c922-402a-9251-10dc0da8c87d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.183164] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a7cb0fd8-040c-4daa-907f-4e312118d1f3 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Lock "2cb0fee4-c922-402a-9251-10dc0da8c87d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.183264] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a7cb0fd8-040c-4daa-907f-4e312118d1f3 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Lock "2cb0fee4-c922-402a-9251-10dc0da8c87d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.186208] env[62385]: INFO nova.compute.manager [None req-a7cb0fd8-040c-4daa-907f-4e312118d1f3 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Terminating instance [ 1069.187856] env[62385]: DEBUG nova.compute.manager [None req-a7cb0fd8-040c-4daa-907f-4e312118d1f3 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1069.188070] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a7cb0fd8-040c-4daa-907f-4e312118d1f3 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1069.188905] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cedba8ca-672d-4570-876d-7327b9812a8f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.197508] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7cb0fd8-040c-4daa-907f-4e312118d1f3 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1069.197743] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fc50b3d3-5ec2-4d61-ba57-8b23fb8f7ef4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.204287] env[62385]: DEBUG oslo_vmware.api [None req-a7cb0fd8-040c-4daa-907f-4e312118d1f3 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Waiting for the task: (returnval){ [ 1069.204287] env[62385]: value = "task-1206458" [ 1069.204287] env[62385]: _type = "Task" [ 1069.204287] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.211762] env[62385]: DEBUG oslo_vmware.api [None req-a7cb0fd8-040c-4daa-907f-4e312118d1f3 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Task: {'id': task-1206458, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.468945] env[62385]: DEBUG oslo_vmware.api [None req-af1917ad-a569-4b83-b638-ab28fd00865a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206452, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.627153] env[62385]: DEBUG nova.network.neutron [-] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1069.649525] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9472d516-b235-4d7f-b241-7bc87bdf00b5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.658524] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4a1ae31-4b37-4766-b920-609ae4e73f84 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.687998] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62385) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1069.688213] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.318s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.688549] env[62385]: DEBUG nova.compute.manager [req-6d33e511-42c9-41ed-8e25-ff29daf9fe69 req-61599d73-67cc-4f24-b5d8-9b0b434383e8 service nova] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Detach interface failed, port_id=dd3797d4-87fb-4629-8b74-115e823503bb, reason: Instance f93d9f5d-f7b6-4988-af7c-877a9e4c2104 could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1069.688962] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1069.689106] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Cleaning up deleted instances {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1069.715310] env[62385]: DEBUG oslo_vmware.api [None req-a7cb0fd8-040c-4daa-907f-4e312118d1f3 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Task: {'id': task-1206458, 'name': PowerOffVM_Task, 'duration_secs': 0.180777} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.715569] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7cb0fd8-040c-4daa-907f-4e312118d1f3 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1069.715739] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a7cb0fd8-040c-4daa-907f-4e312118d1f3 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1069.716030] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a290c955-d80b-4a9e-92c1-3fda1d839370 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.787267] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a7cb0fd8-040c-4daa-907f-4e312118d1f3 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1069.787503] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a7cb0fd8-040c-4daa-907f-4e312118d1f3 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Deleting contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1069.787694] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7cb0fd8-040c-4daa-907f-4e312118d1f3 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Deleting the datastore file [datastore2] 2cb0fee4-c922-402a-9251-10dc0da8c87d {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1069.787966] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3cb8e0fc-4058-4977-b219-c36b2706aa73 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.796297] env[62385]: DEBUG oslo_vmware.api [None req-a7cb0fd8-040c-4daa-907f-4e312118d1f3 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Waiting for the task: (returnval){ [ 1069.796297] env[62385]: value = "task-1206461" [ 1069.796297] env[62385]: _type = "Task" [ 1069.796297] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.804723] env[62385]: DEBUG oslo_vmware.api [None req-a7cb0fd8-040c-4daa-907f-4e312118d1f3 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Task: {'id': task-1206461, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.970361] env[62385]: DEBUG oslo_vmware.api [None req-af1917ad-a569-4b83-b638-ab28fd00865a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206452, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.129138] env[62385]: INFO nova.compute.manager [-] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Took 1.22 seconds to deallocate network for instance. [ 1070.206135] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] There are 60 instances to clean {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1070.206424] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 47311a9c-74d2-400d-a22d-9ef55b14edb5] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1070.305846] env[62385]: DEBUG oslo_vmware.api [None req-a7cb0fd8-040c-4daa-907f-4e312118d1f3 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Task: {'id': task-1206461, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.185248} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.306119] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7cb0fd8-040c-4daa-907f-4e312118d1f3 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1070.306311] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a7cb0fd8-040c-4daa-907f-4e312118d1f3 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Deleted contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1070.306503] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a7cb0fd8-040c-4daa-907f-4e312118d1f3 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1070.306678] env[62385]: INFO nova.compute.manager [None req-a7cb0fd8-040c-4daa-907f-4e312118d1f3 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1070.306916] env[62385]: DEBUG oslo.service.loopingcall [None req-a7cb0fd8-040c-4daa-907f-4e312118d1f3 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1070.307126] env[62385]: DEBUG nova.compute.manager [-] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1070.307242] env[62385]: DEBUG nova.network.neutron [-] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1070.470901] env[62385]: DEBUG oslo_vmware.api [None req-af1917ad-a569-4b83-b638-ab28fd00865a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206452, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.635940] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0a480391-987b-4664-8241-e43a8295ee9e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.636232] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0a480391-987b-4664-8241-e43a8295ee9e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.636462] env[62385]: DEBUG nova.objects.instance [None req-0a480391-987b-4664-8241-e43a8295ee9e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Lazy-loading 'resources' on Instance uuid f93d9f5d-f7b6-4988-af7c-877a9e4c2104 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1070.709433] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: b0a6a3b9-05e4-4190-9cc6-4237dee07f91] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1070.971589] env[62385]: DEBUG oslo_vmware.api [None req-af1917ad-a569-4b83-b638-ab28fd00865a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206452, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.045578] env[62385]: DEBUG nova.network.neutron [-] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1071.177322] env[62385]: DEBUG nova.compute.manager [req-0646a75c-b949-46f6-9ae6-bf1226d3d592 req-fc5b9049-e9e0-46eb-aa8c-b99ae5675a17 service nova] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Received event network-vif-deleted-bc49aae7-4b0a-4cef-9b58-1baefe8a64ce {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1071.212086] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 784589bb-a668-4674-83b8-6219def15067] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1071.279704] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf5ae77e-d50b-4443-a490-4df9d0dc8ed8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.287334] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fd25028-3db5-43b4-8b8d-03b166457a36 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.318259] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95e0fafa-7948-49af-91ad-0d8ad049cb4a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.325942] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85d18c19-1c99-4acc-ae7f-a8bb01a2800a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.338908] env[62385]: DEBUG nova.compute.provider_tree [None req-0a480391-987b-4664-8241-e43a8295ee9e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1071.472371] env[62385]: DEBUG oslo_vmware.api [None req-af1917ad-a569-4b83-b638-ab28fd00865a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206452, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.548572] env[62385]: INFO nova.compute.manager [-] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Took 1.24 seconds to deallocate network for instance. [ 1071.717601] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 76672404-2bcf-438a-9e21-92ea9dc86461] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1071.842392] env[62385]: DEBUG nova.scheduler.client.report [None req-0a480391-987b-4664-8241-e43a8295ee9e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1071.972690] env[62385]: DEBUG oslo_vmware.api [None req-af1917ad-a569-4b83-b638-ab28fd00865a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206452, 'name': ReconfigVM_Task, 'duration_secs': 5.726959} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.972945] env[62385]: DEBUG oslo_concurrency.lockutils [None req-af1917ad-a569-4b83-b638-ab28fd00865a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Releasing lock "68748d25-017f-4b58-9984-70264abbc5ba" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.973193] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-af1917ad-a569-4b83-b638-ab28fd00865a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Reconfigured VM to detach interface {{(pid=62385) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 1072.055191] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a7cb0fd8-040c-4daa-907f-4e312118d1f3 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.220737] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: a1c71d8b-0953-4cc2-961d-0c2b2fc11110] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1072.347840] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0a480391-987b-4664-8241-e43a8295ee9e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.711s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.350255] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a7cb0fd8-040c-4daa-907f-4e312118d1f3 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.295s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.350501] env[62385]: DEBUG nova.objects.instance [None req-a7cb0fd8-040c-4daa-907f-4e312118d1f3 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Lazy-loading 'resources' on Instance uuid 2cb0fee4-c922-402a-9251-10dc0da8c87d {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1072.374893] env[62385]: INFO nova.scheduler.client.report [None req-0a480391-987b-4664-8241-e43a8295ee9e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Deleted allocations for instance f93d9f5d-f7b6-4988-af7c-877a9e4c2104 [ 1072.726463] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 9741e9c8-60d9-41bf-8185-ef3cfcc1c844] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1072.881619] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0a480391-987b-4664-8241-e43a8295ee9e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Lock "f93d9f5d-f7b6-4988-af7c-877a9e4c2104" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.100s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.963303] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6982ef04-20fd-4305-a6ba-223723a5adb6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.970736] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7184733d-8f96-4644-bbbd-c270930df22f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.000458] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a39056e5-2e60-4fd2-9f15-6955f7b3c268 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.007529] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f18ba3b-7d12-4619-981e-5c3147edd600 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.021115] env[62385]: INFO nova.compute.manager [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Rebuilding instance [ 1073.023097] env[62385]: DEBUG nova.compute.provider_tree [None req-a7cb0fd8-040c-4daa-907f-4e312118d1f3 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1073.058327] env[62385]: DEBUG nova.compute.manager [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1073.059198] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-807eab45-69b7-43c6-91df-e17a91153efa {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.230536] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: b1e0baf3-643c-49c2-8a80-8de07d8527e7] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1073.288426] env[62385]: DEBUG oslo_concurrency.lockutils [None req-af1917ad-a569-4b83-b638-ab28fd00865a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "refresh_cache-68748d25-017f-4b58-9984-70264abbc5ba" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1073.288550] env[62385]: DEBUG oslo_concurrency.lockutils [None req-af1917ad-a569-4b83-b638-ab28fd00865a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquired lock "refresh_cache-68748d25-017f-4b58-9984-70264abbc5ba" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1073.288663] env[62385]: DEBUG nova.network.neutron [None req-af1917ad-a569-4b83-b638-ab28fd00865a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1073.525682] env[62385]: DEBUG nova.scheduler.client.report [None req-a7cb0fd8-040c-4daa-907f-4e312118d1f3 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1073.570777] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1073.571151] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b800c20b-8297-47e5-af4c-5bf5b2c35a04 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.578849] env[62385]: DEBUG oslo_vmware.api [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Waiting for the task: (returnval){ [ 1073.578849] env[62385]: value = "task-1206463" [ 1073.578849] env[62385]: _type = "Task" [ 1073.578849] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.586342] env[62385]: DEBUG oslo_vmware.api [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Task: {'id': task-1206463, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.734766] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: df020d6b-3fab-4599-a342-47c7833b4240] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1073.753150] env[62385]: DEBUG oslo_concurrency.lockutils [None req-22768a94-d0a0-4b4d-a916-4455fe2133b2 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "68748d25-017f-4b58-9984-70264abbc5ba" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.753579] env[62385]: DEBUG oslo_concurrency.lockutils [None req-22768a94-d0a0-4b4d-a916-4455fe2133b2 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "68748d25-017f-4b58-9984-70264abbc5ba" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.753870] env[62385]: DEBUG oslo_concurrency.lockutils [None req-22768a94-d0a0-4b4d-a916-4455fe2133b2 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "68748d25-017f-4b58-9984-70264abbc5ba-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.754115] env[62385]: DEBUG oslo_concurrency.lockutils [None req-22768a94-d0a0-4b4d-a916-4455fe2133b2 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "68748d25-017f-4b58-9984-70264abbc5ba-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.754327] env[62385]: DEBUG oslo_concurrency.lockutils [None req-22768a94-d0a0-4b4d-a916-4455fe2133b2 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "68748d25-017f-4b58-9984-70264abbc5ba-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.756866] env[62385]: INFO nova.compute.manager [None req-22768a94-d0a0-4b4d-a916-4455fe2133b2 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Terminating instance [ 1073.759160] env[62385]: DEBUG nova.compute.manager [None req-22768a94-d0a0-4b4d-a916-4455fe2133b2 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1073.759370] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-22768a94-d0a0-4b4d-a916-4455fe2133b2 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1073.760399] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fef6c44-8e89-4315-8f44-699c43023ad3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.769942] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-22768a94-d0a0-4b4d-a916-4455fe2133b2 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1073.770234] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3d90d3b5-248b-4edd-ba95-8c2c47be8138 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.776645] env[62385]: DEBUG oslo_vmware.api [None req-22768a94-d0a0-4b4d-a916-4455fe2133b2 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 1073.776645] env[62385]: value = "task-1206465" [ 1073.776645] env[62385]: _type = "Task" [ 1073.776645] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.785546] env[62385]: DEBUG oslo_vmware.api [None req-22768a94-d0a0-4b4d-a916-4455fe2133b2 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206465, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.018069] env[62385]: INFO nova.network.neutron [None req-af1917ad-a569-4b83-b638-ab28fd00865a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Port 331b6188-b103-437e-8436-5259bc356e31 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1074.018454] env[62385]: DEBUG nova.network.neutron [None req-af1917ad-a569-4b83-b638-ab28fd00865a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Updating instance_info_cache with network_info: [{"id": "c308701b-9b9e-4ff3-9a01-c27aabf54f0d", "address": "fa:16:3e:9a:ed:8c", "network": {"id": "daad7070-dc06-4580-9f4a-ac8aaa546955", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-770473557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.210", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "012426f098ce40c3aaa00f628fe9cebb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "685b4083-b748-41fb-a68a-273b1073fa28", "external-id": "nsx-vlan-transportzone-312", "segmentation_id": 312, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc308701b-9b", "ovs_interfaceid": "c308701b-9b9e-4ff3-9a01-c27aabf54f0d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1074.029801] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a7cb0fd8-040c-4daa-907f-4e312118d1f3 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.680s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.052723] env[62385]: INFO nova.scheduler.client.report [None req-a7cb0fd8-040c-4daa-907f-4e312118d1f3 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Deleted allocations for instance 2cb0fee4-c922-402a-9251-10dc0da8c87d [ 1074.088685] env[62385]: DEBUG oslo_vmware.api [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Task: {'id': task-1206463, 'name': PowerOffVM_Task, 'duration_secs': 0.220802} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.088963] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1074.089670] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1074.089933] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-07ca9893-1e41-4d2b-8cdc-b883142d10bf {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.096120] env[62385]: DEBUG oslo_vmware.api [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Waiting for the task: (returnval){ [ 1074.096120] env[62385]: value = "task-1206466" [ 1074.096120] env[62385]: _type = "Task" [ 1074.096120] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.103628] env[62385]: DEBUG oslo_vmware.api [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Task: {'id': task-1206466, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.238273] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: fd0b59c2-e38d-4ecc-b32f-72f07d555ffd] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1074.286735] env[62385]: DEBUG oslo_vmware.api [None req-22768a94-d0a0-4b4d-a916-4455fe2133b2 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206465, 'name': PowerOffVM_Task, 'duration_secs': 0.193737} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.286939] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-22768a94-d0a0-4b4d-a916-4455fe2133b2 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1074.286978] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-22768a94-d0a0-4b4d-a916-4455fe2133b2 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1074.287249] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fc1f01cb-4fb4-40f2-be9b-8909b9f30374 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.521487] env[62385]: DEBUG oslo_concurrency.lockutils [None req-af1917ad-a569-4b83-b638-ab28fd00865a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Releasing lock "refresh_cache-68748d25-017f-4b58-9984-70264abbc5ba" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1074.559738] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a7cb0fd8-040c-4daa-907f-4e312118d1f3 tempest-ServersNegativeTestMultiTenantJSON-41131069 tempest-ServersNegativeTestMultiTenantJSON-41131069-project-member] Lock "2cb0fee4-c922-402a-9251-10dc0da8c87d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.377s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.607143] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] VM already powered off {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1074.608452] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Volume detach. Driver type: vmdk {{(pid=62385) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1074.608452] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261288', 'volume_id': 'e44d7ff8-cd5f-4a36-a788-a75683085f24', 'name': 'volume-e44d7ff8-cd5f-4a36-a788-a75683085f24', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5a5bcb7c-86cd-4950-a3dd-15515ae0aca6', 'attached_at': '', 'detached_at': '', 'volume_id': 'e44d7ff8-cd5f-4a36-a788-a75683085f24', 'serial': 'e44d7ff8-cd5f-4a36-a788-a75683085f24'} {{(pid=62385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1074.608452] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b966bc4-179c-451c-a637-cb9e7ca1be11 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.627068] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28528fa8-413e-4757-8ea5-5fca62955eaa {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.635403] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc293d20-dd7a-4289-bd36-b6f511a57924 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.657026] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dee629b2-4729-48f7-8ff4-c2893e7fab55 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.668757] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] The volume has not been displaced from its original location: [datastore2] volume-e44d7ff8-cd5f-4a36-a788-a75683085f24/volume-e44d7ff8-cd5f-4a36-a788-a75683085f24.vmdk. No consolidation needed. {{(pid=62385) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1074.674136] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Reconfiguring VM instance instance-0000006a to detach disk 2000 {{(pid=62385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1074.674602] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-547ad128-250b-4e69-b396-891d91469ec3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.694148] env[62385]: DEBUG oslo_vmware.api [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Waiting for the task: (returnval){ [ 1074.694148] env[62385]: value = "task-1206468" [ 1074.694148] env[62385]: _type = "Task" [ 1074.694148] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.702145] env[62385]: DEBUG oslo_vmware.api [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Task: {'id': task-1206468, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.741225] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 104dea2c-6977-4683-bc32-05131858f562] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1075.025274] env[62385]: DEBUG oslo_concurrency.lockutils [None req-af1917ad-a569-4b83-b638-ab28fd00865a tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "interface-68748d25-017f-4b58-9984-70264abbc5ba-331b6188-b103-437e-8436-5259bc356e31" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.633s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.204877] env[62385]: DEBUG oslo_vmware.api [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Task: {'id': task-1206468, 'name': ReconfigVM_Task, 'duration_secs': 0.215824} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.205365] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Reconfigured VM instance instance-0000006a to detach disk 2000 {{(pid=62385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1075.210530] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9b4f4e8e-df0e-4082-ba0a-60de7f6148ae {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.225561] env[62385]: DEBUG oslo_vmware.api [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Waiting for the task: (returnval){ [ 1075.225561] env[62385]: value = "task-1206469" [ 1075.225561] env[62385]: _type = "Task" [ 1075.225561] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.233553] env[62385]: DEBUG oslo_vmware.api [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Task: {'id': task-1206469, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.244191] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 5daf0a99-0c2a-40d8-afc7-1998e21e32d8] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1075.736128] env[62385]: DEBUG oslo_vmware.api [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Task: {'id': task-1206469, 'name': ReconfigVM_Task, 'duration_secs': 0.111722} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.736451] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261288', 'volume_id': 'e44d7ff8-cd5f-4a36-a788-a75683085f24', 'name': 'volume-e44d7ff8-cd5f-4a36-a788-a75683085f24', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5a5bcb7c-86cd-4950-a3dd-15515ae0aca6', 'attached_at': '', 'detached_at': '', 'volume_id': 'e44d7ff8-cd5f-4a36-a788-a75683085f24', 'serial': 'e44d7ff8-cd5f-4a36-a788-a75683085f24'} {{(pid=62385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1075.736747] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1075.737613] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b95fe0a4-cce2-4365-aa47-c18669900aa6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.747251] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: df559cb8-555c-496b-bce3-0981be3ef65f] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1075.750378] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1075.754024] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7a8faf22-d984-4ad9-8e94-53d4446cb6c4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.819588] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1075.819831] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Deleting contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1075.820409] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Deleting the datastore file [datastore2] 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1075.820409] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0eb513e4-be9f-405b-8ba9-34f649b1cdc7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.827739] env[62385]: DEBUG oslo_vmware.api [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Waiting for the task: (returnval){ [ 1075.827739] env[62385]: value = "task-1206471" [ 1075.827739] env[62385]: _type = "Task" [ 1075.827739] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.835977] env[62385]: DEBUG oslo_vmware.api [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Task: {'id': task-1206471, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.252131] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 75cf7907-d9e3-4f54-90c9-f8d714e1df40] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1076.338397] env[62385]: DEBUG oslo_vmware.api [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Task: {'id': task-1206471, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.079364} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.339484] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1076.339484] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Deleted contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1076.339484] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1076.410668] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Volume detach. Driver type: vmdk {{(pid=62385) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1076.411867] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-529da18d-a6b9-430d-8d8d-55b1886c886f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.415156] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-22768a94-d0a0-4b4d-a916-4455fe2133b2 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1076.415361] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-22768a94-d0a0-4b4d-a916-4455fe2133b2 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1076.415541] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-22768a94-d0a0-4b4d-a916-4455fe2133b2 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Deleting the datastore file [datastore1] 68748d25-017f-4b58-9984-70264abbc5ba {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1076.416440] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6a13081f-35df-4d6b-aec4-793eb4ae97e8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.426912] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aff07a7-1a85-4c9b-bfaf-ffc98d20dcde {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.439256] env[62385]: DEBUG oslo_vmware.api [None req-22768a94-d0a0-4b4d-a916-4455fe2133b2 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 1076.439256] env[62385]: value = "task-1206472" [ 1076.439256] env[62385]: _type = "Task" [ 1076.439256] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.447017] env[62385]: DEBUG oslo_vmware.api [None req-22768a94-d0a0-4b4d-a916-4455fe2133b2 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206472, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.456820] env[62385]: ERROR nova.compute.manager [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Failed to detach volume e44d7ff8-cd5f-4a36-a788-a75683085f24 from /dev/sda: nova.exception.InstanceNotFound: Instance 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6 could not be found. [ 1076.456820] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Traceback (most recent call last): [ 1076.456820] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 1076.456820] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] self.driver.rebuild(**kwargs) [ 1076.456820] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 1076.456820] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] raise NotImplementedError() [ 1076.456820] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] NotImplementedError [ 1076.456820] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] [ 1076.456820] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] During handling of the above exception, another exception occurred: [ 1076.456820] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] [ 1076.456820] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Traceback (most recent call last): [ 1076.456820] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 1076.456820] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] self.driver.detach_volume(context, old_connection_info, [ 1076.456820] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 553, in detach_volume [ 1076.456820] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] return self._volumeops.detach_volume(connection_info, instance) [ 1076.456820] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 1076.456820] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] self._detach_volume_vmdk(connection_info, instance) [ 1076.456820] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 1076.456820] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 1076.456820] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1135, in get_vm_ref [ 1076.456820] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] stable_ref.fetch_moref(session) [ 1076.456820] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1126, in fetch_moref [ 1076.456820] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] raise exception.InstanceNotFound(instance_id=self._uuid) [ 1076.456820] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] nova.exception.InstanceNotFound: Instance 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6 could not be found. [ 1076.456820] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] [ 1076.587800] env[62385]: DEBUG nova.compute.utils [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Build of instance 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6 aborted: Failed to rebuild volume backed instance. {{(pid=62385) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 1076.590858] env[62385]: ERROR nova.compute.manager [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6 aborted: Failed to rebuild volume backed instance. [ 1076.590858] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Traceback (most recent call last): [ 1076.590858] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 1076.590858] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] self.driver.rebuild(**kwargs) [ 1076.590858] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 1076.590858] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] raise NotImplementedError() [ 1076.590858] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] NotImplementedError [ 1076.590858] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] [ 1076.590858] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] During handling of the above exception, another exception occurred: [ 1076.590858] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] [ 1076.590858] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Traceback (most recent call last): [ 1076.590858] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] File "/opt/stack/nova/nova/compute/manager.py", line 3600, in _rebuild_volume_backed_instance [ 1076.590858] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] self._detach_root_volume(context, instance, root_bdm) [ 1076.590858] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] File "/opt/stack/nova/nova/compute/manager.py", line 3579, in _detach_root_volume [ 1076.590858] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] with excutils.save_and_reraise_exception(): [ 1076.590858] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1076.590858] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] self.force_reraise() [ 1076.590858] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1076.590858] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] raise self.value [ 1076.590858] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 1076.590858] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] self.driver.detach_volume(context, old_connection_info, [ 1076.590858] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 553, in detach_volume [ 1076.590858] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] return self._volumeops.detach_volume(connection_info, instance) [ 1076.590858] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 1076.590858] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] self._detach_volume_vmdk(connection_info, instance) [ 1076.590858] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 1076.590858] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 1076.590858] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1135, in get_vm_ref [ 1076.590858] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] stable_ref.fetch_moref(session) [ 1076.590858] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1126, in fetch_moref [ 1076.590858] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] raise exception.InstanceNotFound(instance_id=self._uuid) [ 1076.590858] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] nova.exception.InstanceNotFound: Instance 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6 could not be found. [ 1076.590858] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] [ 1076.590858] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] During handling of the above exception, another exception occurred: [ 1076.590858] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] [ 1076.590858] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Traceback (most recent call last): [ 1076.590858] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] File "/opt/stack/nova/nova/compute/manager.py", line 10865, in _error_out_instance_on_exception [ 1076.590858] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] yield [ 1076.590858] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] File "/opt/stack/nova/nova/compute/manager.py", line 3868, in rebuild_instance [ 1076.590858] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] self._do_rebuild_instance_with_claim( [ 1076.592019] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] File "/opt/stack/nova/nova/compute/manager.py", line 3954, in _do_rebuild_instance_with_claim [ 1076.592019] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] self._do_rebuild_instance( [ 1076.592019] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] File "/opt/stack/nova/nova/compute/manager.py", line 4146, in _do_rebuild_instance [ 1076.592019] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] self._rebuild_default_impl(**kwargs) [ 1076.592019] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] File "/opt/stack/nova/nova/compute/manager.py", line 3723, in _rebuild_default_impl [ 1076.592019] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] self._rebuild_volume_backed_instance( [ 1076.592019] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] File "/opt/stack/nova/nova/compute/manager.py", line 3615, in _rebuild_volume_backed_instance [ 1076.592019] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] raise exception.BuildAbortException( [ 1076.592019] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] nova.exception.BuildAbortException: Build of instance 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6 aborted: Failed to rebuild volume backed instance. [ 1076.592019] env[62385]: ERROR nova.compute.manager [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] [ 1076.755842] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: fad3b054-cf2a-4ee3-a2a5-918a96a4a125] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1076.949491] env[62385]: DEBUG oslo_vmware.api [None req-22768a94-d0a0-4b4d-a916-4455fe2133b2 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206472, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.161639} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.949555] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-22768a94-d0a0-4b4d-a916-4455fe2133b2 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1076.949747] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-22768a94-d0a0-4b4d-a916-4455fe2133b2 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1076.949914] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-22768a94-d0a0-4b4d-a916-4455fe2133b2 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1076.950618] env[62385]: INFO nova.compute.manager [None req-22768a94-d0a0-4b4d-a916-4455fe2133b2 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Took 3.19 seconds to destroy the instance on the hypervisor. [ 1076.950618] env[62385]: DEBUG oslo.service.loopingcall [None req-22768a94-d0a0-4b4d-a916-4455fe2133b2 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1076.950618] env[62385]: DEBUG nova.compute.manager [-] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1076.950618] env[62385]: DEBUG nova.network.neutron [-] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1076.983818] env[62385]: DEBUG oslo_concurrency.lockutils [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "c4f0482d-5b46-4c28-9c40-a59b6aee7dc3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.984622] env[62385]: DEBUG oslo_concurrency.lockutils [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "c4f0482d-5b46-4c28-9c40-a59b6aee7dc3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.259719] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: caab700b-bd3c-497e-b40a-eb9d20dfc1e2] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1077.486434] env[62385]: DEBUG nova.compute.manager [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1077.765302] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 09c017a9-ad18-49b6-b72d-958023c81b24] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1077.794124] env[62385]: DEBUG nova.compute.manager [req-5e25d599-500f-4ffa-8acc-e28a90394948 req-84dbbfb9-49a1-49eb-830d-8c38c7999e9c service nova] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Received event network-vif-deleted-c308701b-9b9e-4ff3-9a01-c27aabf54f0d {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1077.794346] env[62385]: INFO nova.compute.manager [req-5e25d599-500f-4ffa-8acc-e28a90394948 req-84dbbfb9-49a1-49eb-830d-8c38c7999e9c service nova] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Neutron deleted interface c308701b-9b9e-4ff3-9a01-c27aabf54f0d; detaching it from the instance and deleting it from the info cache [ 1077.794526] env[62385]: DEBUG nova.network.neutron [req-5e25d599-500f-4ffa-8acc-e28a90394948 req-84dbbfb9-49a1-49eb-830d-8c38c7999e9c service nova] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1078.016727] env[62385]: DEBUG oslo_concurrency.lockutils [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.016727] env[62385]: DEBUG oslo_concurrency.lockutils [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.016727] env[62385]: INFO nova.compute.claims [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1078.264995] env[62385]: DEBUG nova.network.neutron [-] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1078.268107] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 9b2f8292-9d89-407e-96c5-195ee398cc7b] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1078.298073] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-406d6e24-5be6-4a93-9e94-716463bee2fe {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.308481] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70f5f985-69df-4a4d-b503-f94f3f3e4123 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.339162] env[62385]: DEBUG nova.compute.manager [req-5e25d599-500f-4ffa-8acc-e28a90394948 req-84dbbfb9-49a1-49eb-830d-8c38c7999e9c service nova] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Detach interface failed, port_id=c308701b-9b9e-4ff3-9a01-c27aabf54f0d, reason: Instance 68748d25-017f-4b58-9984-70264abbc5ba could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1078.607539] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.770695] env[62385]: INFO nova.compute.manager [-] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Took 1.82 seconds to deallocate network for instance. [ 1078.772781] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: f937ef7b-ab39-4d8a-9577-7faabc652aa5] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1079.124621] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1587f852-ccfc-4af7-b9b2-fa29355d9754 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Acquiring lock "5a5bcb7c-86cd-4950-a3dd-15515ae0aca6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1079.124876] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1587f852-ccfc-4af7-b9b2-fa29355d9754 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Lock "5a5bcb7c-86cd-4950-a3dd-15515ae0aca6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1079.125109] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1587f852-ccfc-4af7-b9b2-fa29355d9754 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Acquiring lock "5a5bcb7c-86cd-4950-a3dd-15515ae0aca6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1079.125727] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1587f852-ccfc-4af7-b9b2-fa29355d9754 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Lock "5a5bcb7c-86cd-4950-a3dd-15515ae0aca6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1079.125921] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1587f852-ccfc-4af7-b9b2-fa29355d9754 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Lock "5a5bcb7c-86cd-4950-a3dd-15515ae0aca6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.127991] env[62385]: INFO nova.compute.manager [None req-1587f852-ccfc-4af7-b9b2-fa29355d9754 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Terminating instance [ 1079.129986] env[62385]: DEBUG nova.compute.manager [None req-1587f852-ccfc-4af7-b9b2-fa29355d9754 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1079.130294] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c1ef64ee-6bd8-418a-95b7-c4abaa2eb4ae {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.133600] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-496f5bf7-e449-4d1a-ac5f-e92c75d7019d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.142877] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5b70c6f-4a5c-4391-904d-b8e167ba8162 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.148632] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2cd4402-69ab-408a-8e58-701d41eb943c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.194287] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9707db99-1475-49a3-a9ea-7a66f38924e5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.197755] env[62385]: WARNING nova.virt.vmwareapi.driver [None req-1587f852-ccfc-4af7-b9b2-fa29355d9754 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6 could not be found. [ 1079.197969] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-1587f852-ccfc-4af7-b9b2-fa29355d9754 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1079.198336] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-db811c99-ec2f-4603-a5ba-1c40a0aba8f6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.206302] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ea8f866-3060-4d2e-9c7e-4f10ff47b4d2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.214322] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3210d298-5c2c-4729-8e7b-c237a09b795e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.241898] env[62385]: DEBUG nova.compute.provider_tree [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1079.252565] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-1587f852-ccfc-4af7-b9b2-fa29355d9754 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6 could not be found. [ 1079.252761] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-1587f852-ccfc-4af7-b9b2-fa29355d9754 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1079.252863] env[62385]: INFO nova.compute.manager [None req-1587f852-ccfc-4af7-b9b2-fa29355d9754 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Took 0.12 seconds to destroy the instance on the hypervisor. [ 1079.253521] env[62385]: DEBUG oslo.service.loopingcall [None req-1587f852-ccfc-4af7-b9b2-fa29355d9754 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1079.254300] env[62385]: DEBUG nova.compute.manager [-] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1079.254400] env[62385]: DEBUG nova.network.neutron [-] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1079.277601] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 2bcd9457-8a73-4e7a-b778-d52c468b3aae] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1079.281369] env[62385]: DEBUG oslo_concurrency.lockutils [None req-22768a94-d0a0-4b4d-a916-4455fe2133b2 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1079.755407] env[62385]: DEBUG nova.scheduler.client.report [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1079.781809] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 4f00d9a4-927e-4f4c-b0b7-de8cc5e61c66] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1079.837935] env[62385]: DEBUG nova.compute.manager [req-5976e21d-df30-41ca-b860-0180d0b92313 req-5b71dfa4-f190-4ab6-9378-35ec9eb03fd3 service nova] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Received event network-vif-deleted-5af3050a-7535-42c3-bb8d-4f1141049e32 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1079.837935] env[62385]: INFO nova.compute.manager [req-5976e21d-df30-41ca-b860-0180d0b92313 req-5b71dfa4-f190-4ab6-9378-35ec9eb03fd3 service nova] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Neutron deleted interface 5af3050a-7535-42c3-bb8d-4f1141049e32; detaching it from the instance and deleting it from the info cache [ 1079.837935] env[62385]: DEBUG nova.network.neutron [req-5976e21d-df30-41ca-b860-0180d0b92313 req-5b71dfa4-f190-4ab6-9378-35ec9eb03fd3 service nova] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1080.208531] env[62385]: DEBUG nova.network.neutron [-] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1080.259997] env[62385]: DEBUG oslo_concurrency.lockutils [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.245s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.260526] env[62385]: DEBUG nova.compute.manager [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1080.263330] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.657s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.285181] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: e9bf98d1-089a-4078-b2f8-eb77943731e5] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1080.340998] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-123c4d81-c7e2-4c93-9f3f-43e7c69f5e8c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.351964] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aaa39df-b186-45cb-9b08-97c0393f88cf {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.382829] env[62385]: DEBUG nova.compute.manager [req-5976e21d-df30-41ca-b860-0180d0b92313 req-5b71dfa4-f190-4ab6-9378-35ec9eb03fd3 service nova] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Detach interface failed, port_id=5af3050a-7535-42c3-bb8d-4f1141049e32, reason: Instance 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6 could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1080.419725] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-715e871f-05ca-410e-8589-0f020e1957ed {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.428243] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72223fe8-28e7-433e-b21f-985058f4dc92 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.470983] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-add46c8b-90bb-4bcc-9cdb-dc8a10d73528 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.475822] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6fa9062-b7c4-47ee-a682-8555cb2438cb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.492588] env[62385]: DEBUG nova.compute.provider_tree [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1080.711602] env[62385]: INFO nova.compute.manager [-] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Took 1.46 seconds to deallocate network for instance. [ 1080.768284] env[62385]: DEBUG nova.compute.utils [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1080.771225] env[62385]: DEBUG nova.compute.manager [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1080.771517] env[62385]: DEBUG nova.network.neutron [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1080.791667] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: a58cb15f-e4df-4b1e-b09c-2f61dce7200f] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1080.827046] env[62385]: DEBUG nova.policy [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b314c392b9ca405480c27c25364fda81', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '93111f5ba17c47de9691f90b99e03aa8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 1080.995588] env[62385]: DEBUG nova.scheduler.client.report [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1081.258465] env[62385]: INFO nova.compute.manager [None req-1587f852-ccfc-4af7-b9b2-fa29355d9754 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Took 0.55 seconds to detach 1 volumes for instance. [ 1081.264021] env[62385]: DEBUG nova.compute.manager [None req-1587f852-ccfc-4af7-b9b2-fa29355d9754 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Deleting volume: e44d7ff8-cd5f-4a36-a788-a75683085f24 {{(pid=62385) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 1081.275501] env[62385]: DEBUG nova.compute.manager [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1081.294992] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: d12ffea4-2275-4974-a5c6-b7fa365865b4] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1081.501114] env[62385]: DEBUG oslo_concurrency.lockutils [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.238s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.501451] env[62385]: INFO nova.compute.manager [None req-c836ad98-e403-41e2-b3b0-f7f3926ba3af tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Successfully reverted task state from rebuilding on failure for instance. [ 1081.509145] env[62385]: DEBUG oslo_concurrency.lockutils [None req-22768a94-d0a0-4b4d-a916-4455fe2133b2 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.228s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1081.509486] env[62385]: DEBUG nova.objects.instance [None req-22768a94-d0a0-4b4d-a916-4455fe2133b2 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lazy-loading 'resources' on Instance uuid 68748d25-017f-4b58-9984-70264abbc5ba {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1081.546313] env[62385]: DEBUG nova.network.neutron [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Successfully created port: 07abbf5c-f2dc-4f07-a1a8-f81f59c6a2ae {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1081.781678] env[62385]: INFO nova.virt.block_device [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Booting with volume 6c3e125e-0b69-44e6-9e1a-3527fedf6e2a at /dev/sda [ 1081.797904] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: bd7263b2-b996-4794-946b-2c28215574cb] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1081.822518] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1587f852-ccfc-4af7-b9b2-fa29355d9754 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.837084] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9d4abf90-c336-4131-8a71-db5dded5cdde {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.848124] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6fe24ab-76ac-4bca-9070-8a2a640d533f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.877463] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-baa25723-53fa-49d0-96ce-1d6a63176ec7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.887012] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0748d875-1931-4ed0-b922-f459eddd6d5c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.919502] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d923168-6f3e-4322-887a-a2e09dec04aa {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.925174] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-615e462f-525c-4c14-8ac1-e9d136c5ff6c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.938246] env[62385]: DEBUG nova.virt.block_device [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Updating existing volume attachment record: 7d66210e-294d-4be3-952b-17c2621eb2a6 {{(pid=62385) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1082.136827] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c36c03b4-3477-4ca7-87e8-cc43a29776c2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.143838] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eff410d-1d2c-48c8-be1a-d7ab9fa60eba {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.175015] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b89e0c21-8ca3-471d-a159-8db88cce7520 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.183497] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67b94932-65b2-4595-8d70-6042f55ff7b0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.197247] env[62385]: DEBUG nova.compute.provider_tree [None req-22768a94-d0a0-4b4d-a916-4455fe2133b2 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1082.302416] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: fe0624a9-09a8-498b-bb3c-fda6cab92341] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1082.700091] env[62385]: DEBUG nova.scheduler.client.report [None req-22768a94-d0a0-4b4d-a916-4455fe2133b2 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1082.808206] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: fbc71ada-d3b5-46f6-90a9-489c118d5126] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1083.064372] env[62385]: DEBUG nova.compute.manager [req-12e8727f-1442-4780-886a-f045b33c77d6 req-1833ea9b-d1d8-40c3-b29a-d4b3aa5e633e service nova] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Received event network-vif-plugged-07abbf5c-f2dc-4f07-a1a8-f81f59c6a2ae {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1083.064372] env[62385]: DEBUG oslo_concurrency.lockutils [req-12e8727f-1442-4780-886a-f045b33c77d6 req-1833ea9b-d1d8-40c3-b29a-d4b3aa5e633e service nova] Acquiring lock "c4f0482d-5b46-4c28-9c40-a59b6aee7dc3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.064372] env[62385]: DEBUG oslo_concurrency.lockutils [req-12e8727f-1442-4780-886a-f045b33c77d6 req-1833ea9b-d1d8-40c3-b29a-d4b3aa5e633e service nova] Lock "c4f0482d-5b46-4c28-9c40-a59b6aee7dc3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1083.064372] env[62385]: DEBUG oslo_concurrency.lockutils [req-12e8727f-1442-4780-886a-f045b33c77d6 req-1833ea9b-d1d8-40c3-b29a-d4b3aa5e633e service nova] Lock "c4f0482d-5b46-4c28-9c40-a59b6aee7dc3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1083.064372] env[62385]: DEBUG nova.compute.manager [req-12e8727f-1442-4780-886a-f045b33c77d6 req-1833ea9b-d1d8-40c3-b29a-d4b3aa5e633e service nova] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] No waiting events found dispatching network-vif-plugged-07abbf5c-f2dc-4f07-a1a8-f81f59c6a2ae {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1083.064372] env[62385]: WARNING nova.compute.manager [req-12e8727f-1442-4780-886a-f045b33c77d6 req-1833ea9b-d1d8-40c3-b29a-d4b3aa5e633e service nova] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Received unexpected event network-vif-plugged-07abbf5c-f2dc-4f07-a1a8-f81f59c6a2ae for instance with vm_state building and task_state block_device_mapping. [ 1083.097378] env[62385]: DEBUG nova.network.neutron [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Successfully updated port: 07abbf5c-f2dc-4f07-a1a8-f81f59c6a2ae {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1083.205164] env[62385]: DEBUG oslo_concurrency.lockutils [None req-22768a94-d0a0-4b4d-a916-4455fe2133b2 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.696s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1083.207730] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1587f852-ccfc-4af7-b9b2-fa29355d9754 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.386s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1083.208341] env[62385]: DEBUG nova.objects.instance [None req-1587f852-ccfc-4af7-b9b2-fa29355d9754 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Lazy-loading 'resources' on Instance uuid 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1083.229071] env[62385]: INFO nova.scheduler.client.report [None req-22768a94-d0a0-4b4d-a916-4455fe2133b2 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Deleted allocations for instance 68748d25-017f-4b58-9984-70264abbc5ba [ 1083.312116] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: b7a286a2-9a19-4ef1-8a46-6e2a04e22413] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1083.599430] env[62385]: DEBUG oslo_concurrency.lockutils [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "refresh_cache-c4f0482d-5b46-4c28-9c40-a59b6aee7dc3" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.599588] env[62385]: DEBUG oslo_concurrency.lockutils [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquired lock "refresh_cache-c4f0482d-5b46-4c28-9c40-a59b6aee7dc3" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.599725] env[62385]: DEBUG nova.network.neutron [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1083.736112] env[62385]: DEBUG oslo_concurrency.lockutils [None req-22768a94-d0a0-4b4d-a916-4455fe2133b2 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "68748d25-017f-4b58-9984-70264abbc5ba" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.983s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1083.801998] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92f28a30-b916-4eaf-9341-99144bc0ac61 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.810429] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f95ee40d-ce0c-4472-a5e4-1defdc89f006 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.814154] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 699a57a9-5a1a-4cd1-8449-723400d9caf4] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1083.842649] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b02f02f-e96c-4eea-8597-ea46480e6df9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.850715] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1907b88-a950-4ae9-b50b-c3ea2c744756 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.864451] env[62385]: DEBUG nova.compute.provider_tree [None req-1587f852-ccfc-4af7-b9b2-fa29355d9754 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1084.036249] env[62385]: DEBUG nova.compute.manager [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1084.036808] env[62385]: DEBUG nova.virt.hardware [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1084.037037] env[62385]: DEBUG nova.virt.hardware [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1084.037208] env[62385]: DEBUG nova.virt.hardware [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1084.037396] env[62385]: DEBUG nova.virt.hardware [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1084.037548] env[62385]: DEBUG nova.virt.hardware [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1084.037700] env[62385]: DEBUG nova.virt.hardware [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1084.038094] env[62385]: DEBUG nova.virt.hardware [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1084.038094] env[62385]: DEBUG nova.virt.hardware [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1084.038267] env[62385]: DEBUG nova.virt.hardware [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1084.038434] env[62385]: DEBUG nova.virt.hardware [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1084.038610] env[62385]: DEBUG nova.virt.hardware [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1084.039887] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cab8e12-bc21-4a4b-8417-88db602af1da {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.048050] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a39843b5-919b-4ee6-b420-5f073a24682c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.129695] env[62385]: DEBUG nova.network.neutron [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1084.270031] env[62385]: DEBUG nova.network.neutron [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Updating instance_info_cache with network_info: [{"id": "07abbf5c-f2dc-4f07-a1a8-f81f59c6a2ae", "address": "fa:16:3e:a4:98:71", "network": {"id": "afae6b07-d08c-413f-aa73-b545db1bb766", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1654965885-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "93111f5ba17c47de9691f90b99e03aa8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap07abbf5c-f2", "ovs_interfaceid": "07abbf5c-f2dc-4f07-a1a8-f81f59c6a2ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1084.342217] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 5e5f43e8-ce45-45c7-a244-bfa724ea5e3c] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1084.370310] env[62385]: DEBUG nova.scheduler.client.report [None req-1587f852-ccfc-4af7-b9b2-fa29355d9754 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1084.772927] env[62385]: DEBUG oslo_concurrency.lockutils [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Releasing lock "refresh_cache-c4f0482d-5b46-4c28-9c40-a59b6aee7dc3" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1084.773265] env[62385]: DEBUG nova.compute.manager [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Instance network_info: |[{"id": "07abbf5c-f2dc-4f07-a1a8-f81f59c6a2ae", "address": "fa:16:3e:a4:98:71", "network": {"id": "afae6b07-d08c-413f-aa73-b545db1bb766", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1654965885-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "93111f5ba17c47de9691f90b99e03aa8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap07abbf5c-f2", "ovs_interfaceid": "07abbf5c-f2dc-4f07-a1a8-f81f59c6a2ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1084.773931] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a4:98:71', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4adc8ed0-d11a-4510-9be0-b27c0da3a903', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '07abbf5c-f2dc-4f07-a1a8-f81f59c6a2ae', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1084.781837] env[62385]: DEBUG oslo.service.loopingcall [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1084.782076] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1084.782301] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-38db81af-6ca1-4a56-bd4d-f69de107b736 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.796822] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8fc31933-dad3-45ab-8818-6c7ed1cffb50 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "57637d6e-39e2-405c-bbcf-64bf0ffbebd1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1084.797076] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8fc31933-dad3-45ab-8818-6c7ed1cffb50 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "57637d6e-39e2-405c-bbcf-64bf0ffbebd1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.797294] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8fc31933-dad3-45ab-8818-6c7ed1cffb50 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "57637d6e-39e2-405c-bbcf-64bf0ffbebd1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1084.797483] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8fc31933-dad3-45ab-8818-6c7ed1cffb50 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "57637d6e-39e2-405c-bbcf-64bf0ffbebd1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.797651] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8fc31933-dad3-45ab-8818-6c7ed1cffb50 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "57637d6e-39e2-405c-bbcf-64bf0ffbebd1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.800091] env[62385]: INFO nova.compute.manager [None req-8fc31933-dad3-45ab-8818-6c7ed1cffb50 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Terminating instance [ 1084.801936] env[62385]: DEBUG nova.compute.manager [None req-8fc31933-dad3-45ab-8818-6c7ed1cffb50 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1084.802155] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8fc31933-dad3-45ab-8818-6c7ed1cffb50 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1084.803451] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb3cade9-a2c0-4ec5-ba10-0f465009e7b6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.808713] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1084.808713] env[62385]: value = "task-1206474" [ 1084.808713] env[62385]: _type = "Task" [ 1084.808713] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.813857] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fc31933-dad3-45ab-8818-6c7ed1cffb50 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1084.814419] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e4a84673-ad62-474b-8de6-7e8019ff4b1f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.818645] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206474, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.823338] env[62385]: DEBUG oslo_vmware.api [None req-8fc31933-dad3-45ab-8818-6c7ed1cffb50 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 1084.823338] env[62385]: value = "task-1206475" [ 1084.823338] env[62385]: _type = "Task" [ 1084.823338] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.830768] env[62385]: DEBUG oslo_vmware.api [None req-8fc31933-dad3-45ab-8818-6c7ed1cffb50 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206475, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.846028] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 2a579d13-5372-4340-b7b3-cc02c1912624] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1084.873943] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1587f852-ccfc-4af7-b9b2-fa29355d9754 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.666s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.085921] env[62385]: DEBUG nova.compute.manager [req-ec071fa9-9180-41e2-8480-e88423f4b639 req-986ec5ed-b6ca-41cb-9d2c-3b1fc7f91ea7 service nova] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Received event network-changed-07abbf5c-f2dc-4f07-a1a8-f81f59c6a2ae {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1085.086165] env[62385]: DEBUG nova.compute.manager [req-ec071fa9-9180-41e2-8480-e88423f4b639 req-986ec5ed-b6ca-41cb-9d2c-3b1fc7f91ea7 service nova] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Refreshing instance network info cache due to event network-changed-07abbf5c-f2dc-4f07-a1a8-f81f59c6a2ae. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1085.086412] env[62385]: DEBUG oslo_concurrency.lockutils [req-ec071fa9-9180-41e2-8480-e88423f4b639 req-986ec5ed-b6ca-41cb-9d2c-3b1fc7f91ea7 service nova] Acquiring lock "refresh_cache-c4f0482d-5b46-4c28-9c40-a59b6aee7dc3" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1085.086593] env[62385]: DEBUG oslo_concurrency.lockutils [req-ec071fa9-9180-41e2-8480-e88423f4b639 req-986ec5ed-b6ca-41cb-9d2c-3b1fc7f91ea7 service nova] Acquired lock "refresh_cache-c4f0482d-5b46-4c28-9c40-a59b6aee7dc3" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1085.086789] env[62385]: DEBUG nova.network.neutron [req-ec071fa9-9180-41e2-8480-e88423f4b639 req-986ec5ed-b6ca-41cb-9d2c-3b1fc7f91ea7 service nova] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Refreshing network info cache for port 07abbf5c-f2dc-4f07-a1a8-f81f59c6a2ae {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1085.229140] env[62385]: DEBUG oslo_concurrency.lockutils [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Acquiring lock "c49d4366-2ea0-44f2-8463-daf62bd1c40c" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1085.229446] env[62385]: DEBUG oslo_concurrency.lockutils [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Lock "c49d4366-2ea0-44f2-8463-daf62bd1c40c" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.229796] env[62385]: INFO nova.compute.manager [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Shelving [ 1085.319131] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206474, 'name': CreateVM_Task, 'duration_secs': 0.327254} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.319298] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1085.319965] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'disk_bus': None, 'attachment_id': '7d66210e-294d-4be3-952b-17c2621eb2a6', 'boot_index': 0, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261311', 'volume_id': '6c3e125e-0b69-44e6-9e1a-3527fedf6e2a', 'name': 'volume-6c3e125e-0b69-44e6-9e1a-3527fedf6e2a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c4f0482d-5b46-4c28-9c40-a59b6aee7dc3', 'attached_at': '', 'detached_at': '', 'volume_id': '6c3e125e-0b69-44e6-9e1a-3527fedf6e2a', 'serial': '6c3e125e-0b69-44e6-9e1a-3527fedf6e2a'}, 'delete_on_termination': True, 'mount_device': '/dev/sda', 'device_type': None, 'guest_format': None, 'volume_type': None}], 'swap': None} {{(pid=62385) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1085.320248] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Root volume attach. Driver type: vmdk {{(pid=62385) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1085.321039] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92ad4c7b-9274-4afb-9782-a22e8bcac770 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.329785] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-257d462a-2b46-4373-8d9a-8090dc699d0a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.334656] env[62385]: DEBUG oslo_vmware.api [None req-8fc31933-dad3-45ab-8818-6c7ed1cffb50 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206475, 'name': PowerOffVM_Task, 'duration_secs': 0.200228} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.335198] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fc31933-dad3-45ab-8818-6c7ed1cffb50 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1085.335388] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8fc31933-dad3-45ab-8818-6c7ed1cffb50 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1085.335617] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-404a1409-eb9d-4bd2-bbe1-1ef9bc7009d6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.338736] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24fb50b0-5a3c-43e6-88e5-b6db08d379ba {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.344206] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-fea525a0-2444-4eb7-9546-e504fb35963a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.349235] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: b73a31df-53c8-4550-bf75-5cf3b5aff86c] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1085.352096] env[62385]: DEBUG oslo_vmware.api [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1085.352096] env[62385]: value = "task-1206477" [ 1085.352096] env[62385]: _type = "Task" [ 1085.352096] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.359875] env[62385]: DEBUG oslo_vmware.api [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206477, 'name': RelocateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.397200] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1587f852-ccfc-4af7-b9b2-fa29355d9754 tempest-ServerActionsV293TestJSON-1732052907 tempest-ServerActionsV293TestJSON-1732052907-project-member] Lock "5a5bcb7c-86cd-4950-a3dd-15515ae0aca6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.272s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.405382] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8fc31933-dad3-45ab-8818-6c7ed1cffb50 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1085.405603] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8fc31933-dad3-45ab-8818-6c7ed1cffb50 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1085.405833] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fc31933-dad3-45ab-8818-6c7ed1cffb50 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Deleting the datastore file [datastore1] 57637d6e-39e2-405c-bbcf-64bf0ffbebd1 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1085.406119] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-277d30c0-bf58-44cc-8743-bc10e6d6f4e2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.412263] env[62385]: DEBUG oslo_vmware.api [None req-8fc31933-dad3-45ab-8818-6c7ed1cffb50 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for the task: (returnval){ [ 1085.412263] env[62385]: value = "task-1206478" [ 1085.412263] env[62385]: _type = "Task" [ 1085.412263] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.422290] env[62385]: DEBUG oslo_vmware.api [None req-8fc31933-dad3-45ab-8818-6c7ed1cffb50 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206478, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.484143] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f8b98a55-f453-48cc-a30e-66e4d31671ca tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquiring lock "8e1d416c-2624-468c-94d7-6265e9f4178f" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1085.484421] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f8b98a55-f453-48cc-a30e-66e4d31671ca tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lock "8e1d416c-2624-468c-94d7-6265e9f4178f" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.737184] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1085.737404] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5796d64c-2a8f-42fa-87b2-43884f580109 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.747649] env[62385]: DEBUG oslo_vmware.api [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for the task: (returnval){ [ 1085.747649] env[62385]: value = "task-1206479" [ 1085.747649] env[62385]: _type = "Task" [ 1085.747649] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.758333] env[62385]: DEBUG oslo_vmware.api [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206479, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.817667] env[62385]: DEBUG nova.network.neutron [req-ec071fa9-9180-41e2-8480-e88423f4b639 req-986ec5ed-b6ca-41cb-9d2c-3b1fc7f91ea7 service nova] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Updated VIF entry in instance network info cache for port 07abbf5c-f2dc-4f07-a1a8-f81f59c6a2ae. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1085.818086] env[62385]: DEBUG nova.network.neutron [req-ec071fa9-9180-41e2-8480-e88423f4b639 req-986ec5ed-b6ca-41cb-9d2c-3b1fc7f91ea7 service nova] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Updating instance_info_cache with network_info: [{"id": "07abbf5c-f2dc-4f07-a1a8-f81f59c6a2ae", "address": "fa:16:3e:a4:98:71", "network": {"id": "afae6b07-d08c-413f-aa73-b545db1bb766", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1654965885-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "93111f5ba17c47de9691f90b99e03aa8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap07abbf5c-f2", "ovs_interfaceid": "07abbf5c-f2dc-4f07-a1a8-f81f59c6a2ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1085.853826] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 7f2fe830-e16d-4684-91a4-aab219468e77] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1085.866759] env[62385]: DEBUG oslo_vmware.api [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206477, 'name': RelocateVM_Task} progress is 43%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.925866] env[62385]: DEBUG oslo_vmware.api [None req-8fc31933-dad3-45ab-8818-6c7ed1cffb50 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Task: {'id': task-1206478, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144185} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.926199] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fc31933-dad3-45ab-8818-6c7ed1cffb50 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1085.926403] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8fc31933-dad3-45ab-8818-6c7ed1cffb50 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1085.926599] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-8fc31933-dad3-45ab-8818-6c7ed1cffb50 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1085.926791] env[62385]: INFO nova.compute.manager [None req-8fc31933-dad3-45ab-8818-6c7ed1cffb50 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1085.927073] env[62385]: DEBUG oslo.service.loopingcall [None req-8fc31933-dad3-45ab-8818-6c7ed1cffb50 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1085.927321] env[62385]: DEBUG nova.compute.manager [-] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1085.927450] env[62385]: DEBUG nova.network.neutron [-] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1085.987949] env[62385]: INFO nova.compute.manager [None req-f8b98a55-f453-48cc-a30e-66e4d31671ca tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Detaching volume 925fe844-0d8f-4668-ac82-328339e88cac [ 1086.031589] env[62385]: INFO nova.virt.block_device [None req-f8b98a55-f453-48cc-a30e-66e4d31671ca tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Attempting to driver detach volume 925fe844-0d8f-4668-ac82-328339e88cac from mountpoint /dev/sdb [ 1086.031873] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8b98a55-f453-48cc-a30e-66e4d31671ca tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Volume detach. Driver type: vmdk {{(pid=62385) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1086.032093] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8b98a55-f453-48cc-a30e-66e4d31671ca tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261292', 'volume_id': '925fe844-0d8f-4668-ac82-328339e88cac', 'name': 'volume-925fe844-0d8f-4668-ac82-328339e88cac', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '8e1d416c-2624-468c-94d7-6265e9f4178f', 'attached_at': '', 'detached_at': '', 'volume_id': '925fe844-0d8f-4668-ac82-328339e88cac', 'serial': '925fe844-0d8f-4668-ac82-328339e88cac'} {{(pid=62385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1086.032987] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-764d3b3b-faae-4d65-b3b1-e0330cf138c2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.057313] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df2d1034-09d3-4b6e-a2d4-95b19d71b12d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.066712] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a969412d-7e39-44ca-85fd-186b2337361a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.089788] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfff72fa-9a12-4278-afe6-a8789fb85f35 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.107198] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8b98a55-f453-48cc-a30e-66e4d31671ca tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] The volume has not been displaced from its original location: [datastore1] volume-925fe844-0d8f-4668-ac82-328339e88cac/volume-925fe844-0d8f-4668-ac82-328339e88cac.vmdk. No consolidation needed. {{(pid=62385) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1086.112589] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8b98a55-f453-48cc-a30e-66e4d31671ca tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Reconfiguring VM instance instance-0000005b to detach disk 2001 {{(pid=62385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1086.112974] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1bb72868-19eb-4253-9e30-b9a5680fe354 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.135695] env[62385]: DEBUG oslo_vmware.api [None req-f8b98a55-f453-48cc-a30e-66e4d31671ca tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1086.135695] env[62385]: value = "task-1206480" [ 1086.135695] env[62385]: _type = "Task" [ 1086.135695] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.148711] env[62385]: DEBUG oslo_vmware.api [None req-f8b98a55-f453-48cc-a30e-66e4d31671ca tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206480, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.261155] env[62385]: DEBUG oslo_vmware.api [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206479, 'name': PowerOffVM_Task, 'duration_secs': 0.268972} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.261447] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1086.262289] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eb3f781-fe00-4984-a61f-5046f2fb042e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.282958] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca07bf0f-f1c5-46da-a500-bcd8ee05555c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.321328] env[62385]: DEBUG oslo_concurrency.lockutils [req-ec071fa9-9180-41e2-8480-e88423f4b639 req-986ec5ed-b6ca-41cb-9d2c-3b1fc7f91ea7 service nova] Releasing lock "refresh_cache-c4f0482d-5b46-4c28-9c40-a59b6aee7dc3" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1086.360891] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 5a82fc98-e656-4476-b7dc-99466e9d6afd] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1086.369870] env[62385]: DEBUG oslo_vmware.api [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206477, 'name': RelocateVM_Task} progress is 56%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.648194] env[62385]: DEBUG oslo_vmware.api [None req-f8b98a55-f453-48cc-a30e-66e4d31671ca tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206480, 'name': ReconfigVM_Task, 'duration_secs': 0.267868} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.648501] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8b98a55-f453-48cc-a30e-66e4d31671ca tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Reconfigured VM instance instance-0000005b to detach disk 2001 {{(pid=62385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1086.653409] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2b1c4ee0-edf1-4ce3-a41a-c6f29e87fc43 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.670787] env[62385]: DEBUG oslo_vmware.api [None req-f8b98a55-f453-48cc-a30e-66e4d31671ca tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1086.670787] env[62385]: value = "task-1206481" [ 1086.670787] env[62385]: _type = "Task" [ 1086.670787] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.681514] env[62385]: DEBUG oslo_vmware.api [None req-f8b98a55-f453-48cc-a30e-66e4d31671ca tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206481, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.694154] env[62385]: DEBUG nova.network.neutron [-] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1086.796613] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Creating Snapshot of the VM instance {{(pid=62385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1086.796918] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-d290158d-51b9-4c61-a333-3fafd2aa1363 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.806705] env[62385]: DEBUG oslo_vmware.api [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for the task: (returnval){ [ 1086.806705] env[62385]: value = "task-1206482" [ 1086.806705] env[62385]: _type = "Task" [ 1086.806705] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.818454] env[62385]: DEBUG oslo_vmware.api [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206482, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.865080] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 679150f5-ad17-45e6-9e3d-9f6e28b1aee3] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1086.870754] env[62385]: DEBUG oslo_vmware.api [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206477, 'name': RelocateVM_Task} progress is 71%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.113932] env[62385]: DEBUG nova.compute.manager [req-67c5be74-00aa-4d75-9e1b-80e300d953a9 req-6fd4990e-6a17-4034-9cf7-4679a19a69ef service nova] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Received event network-vif-deleted-a7153a2d-db69-4d14-8ddc-b634ddabbc18 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1087.182672] env[62385]: DEBUG oslo_vmware.api [None req-f8b98a55-f453-48cc-a30e-66e4d31671ca tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206481, 'name': ReconfigVM_Task, 'duration_secs': 0.146352} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.184063] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8b98a55-f453-48cc-a30e-66e4d31671ca tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261292', 'volume_id': '925fe844-0d8f-4668-ac82-328339e88cac', 'name': 'volume-925fe844-0d8f-4668-ac82-328339e88cac', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '8e1d416c-2624-468c-94d7-6265e9f4178f', 'attached_at': '', 'detached_at': '', 'volume_id': '925fe844-0d8f-4668-ac82-328339e88cac', 'serial': '925fe844-0d8f-4668-ac82-328339e88cac'} {{(pid=62385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1087.197362] env[62385]: INFO nova.compute.manager [-] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Took 1.27 seconds to deallocate network for instance. [ 1087.320686] env[62385]: DEBUG oslo_vmware.api [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206482, 'name': CreateSnapshot_Task, 'duration_secs': 0.47893} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.321162] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Created Snapshot of the VM instance {{(pid=62385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1087.321976] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37327aab-008b-4e6e-889a-dc6d254dc091 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.370773] env[62385]: DEBUG oslo_vmware.api [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206477, 'name': RelocateVM_Task} progress is 86%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.372328] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 9b9fd167-d843-4b9d-9e4f-dbde5d3d629e] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1087.704390] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8fc31933-dad3-45ab-8818-6c7ed1cffb50 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.704672] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8fc31933-dad3-45ab-8818-6c7ed1cffb50 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.704910] env[62385]: DEBUG nova.objects.instance [None req-8fc31933-dad3-45ab-8818-6c7ed1cffb50 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lazy-loading 'resources' on Instance uuid 57637d6e-39e2-405c-bbcf-64bf0ffbebd1 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1087.729165] env[62385]: DEBUG nova.objects.instance [None req-f8b98a55-f453-48cc-a30e-66e4d31671ca tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lazy-loading 'flavor' on Instance uuid 8e1d416c-2624-468c-94d7-6265e9f4178f {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1087.845094] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Creating linked-clone VM from snapshot {{(pid=62385) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1087.845496] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-384601ba-becb-4994-af36-636480ed1ec5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.853376] env[62385]: DEBUG oslo_vmware.api [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for the task: (returnval){ [ 1087.853376] env[62385]: value = "task-1206483" [ 1087.853376] env[62385]: _type = "Task" [ 1087.853376] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.861648] env[62385]: DEBUG oslo_vmware.api [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206483, 'name': CloneVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.870426] env[62385]: DEBUG oslo_vmware.api [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206477, 'name': RelocateVM_Task} progress is 97%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.875432] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 55d62bba-ff30-42bf-b1cb-7567988a9361] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1088.319685] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b6ace30-be36-4417-aae8-968e4b085ad0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.327975] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a5ce0d5-a089-41be-93f4-771c1973d6f5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.361989] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6940a47d-01c7-4736-b080-a94393fba553 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.371133] env[62385]: DEBUG oslo_vmware.api [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206483, 'name': CloneVM_Task} progress is 94%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.374729] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e853b5f2-7fc6-4c02-8eb3-cb6a281acecf {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.382165] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: f2be792c-a723-4a69-b255-bbe7f0e8692d] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1088.384140] env[62385]: DEBUG oslo_vmware.api [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206477, 'name': RelocateVM_Task} progress is 97%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.394366] env[62385]: DEBUG nova.compute.provider_tree [None req-8fc31933-dad3-45ab-8818-6c7ed1cffb50 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1088.737655] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f8b98a55-f453-48cc-a30e-66e4d31671ca tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lock "8e1d416c-2624-468c-94d7-6265e9f4178f" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.252s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.867969] env[62385]: DEBUG oslo_vmware.api [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206483, 'name': CloneVM_Task} progress is 94%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.875417] env[62385]: DEBUG oslo_vmware.api [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206477, 'name': RelocateVM_Task} progress is 97%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.885114] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: e4089e71-eea1-4e16-b90c-97966333f2bc] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1088.896823] env[62385]: DEBUG nova.scheduler.client.report [None req-8fc31933-dad3-45ab-8818-6c7ed1cffb50 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1089.368168] env[62385]: DEBUG oslo_vmware.api [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206483, 'name': CloneVM_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.376064] env[62385]: DEBUG oslo_vmware.api [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206477, 'name': RelocateVM_Task, 'duration_secs': 3.874669} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.376330] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Volume attach. Driver type: vmdk {{(pid=62385) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1089.376534] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261311', 'volume_id': '6c3e125e-0b69-44e6-9e1a-3527fedf6e2a', 'name': 'volume-6c3e125e-0b69-44e6-9e1a-3527fedf6e2a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c4f0482d-5b46-4c28-9c40-a59b6aee7dc3', 'attached_at': '', 'detached_at': '', 'volume_id': '6c3e125e-0b69-44e6-9e1a-3527fedf6e2a', 'serial': '6c3e125e-0b69-44e6-9e1a-3527fedf6e2a'} {{(pid=62385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1089.377332] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5503a48-09a9-48f5-a6b7-39fc6f0d57bb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.392368] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 2b744879-01d2-440e-8adf-58438b922888] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1089.395319] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f923eb0-42da-46a5-bc17-2dd764215f2e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.409838] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8fc31933-dad3-45ab-8818-6c7ed1cffb50 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.705s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.419818] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Reconfiguring VM instance instance-00000070 to attach disk [datastore2] volume-6c3e125e-0b69-44e6-9e1a-3527fedf6e2a/volume-6c3e125e-0b69-44e6-9e1a-3527fedf6e2a.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1089.420775] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-59e4238b-2e67-4431-8cf5-2cad72a9ea8a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.435359] env[62385]: INFO nova.scheduler.client.report [None req-8fc31933-dad3-45ab-8818-6c7ed1cffb50 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Deleted allocations for instance 57637d6e-39e2-405c-bbcf-64bf0ffbebd1 [ 1089.444298] env[62385]: DEBUG oslo_vmware.api [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1089.444298] env[62385]: value = "task-1206484" [ 1089.444298] env[62385]: _type = "Task" [ 1089.444298] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.451659] env[62385]: DEBUG oslo_vmware.api [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206484, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.748117] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b9208fd9-d326-4d44-9687-b1a677c4fbe3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquiring lock "8e1d416c-2624-468c-94d7-6265e9f4178f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.748453] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b9208fd9-d326-4d44-9687-b1a677c4fbe3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lock "8e1d416c-2624-468c-94d7-6265e9f4178f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.748942] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b9208fd9-d326-4d44-9687-b1a677c4fbe3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquiring lock "8e1d416c-2624-468c-94d7-6265e9f4178f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.749076] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b9208fd9-d326-4d44-9687-b1a677c4fbe3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lock "8e1d416c-2624-468c-94d7-6265e9f4178f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.749347] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b9208fd9-d326-4d44-9687-b1a677c4fbe3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lock "8e1d416c-2624-468c-94d7-6265e9f4178f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.751728] env[62385]: INFO nova.compute.manager [None req-b9208fd9-d326-4d44-9687-b1a677c4fbe3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Terminating instance [ 1089.753852] env[62385]: DEBUG nova.compute.manager [None req-b9208fd9-d326-4d44-9687-b1a677c4fbe3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1089.754093] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-b9208fd9-d326-4d44-9687-b1a677c4fbe3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1089.754975] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26b676a3-d78e-48b9-b536-8e109bb239c7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.762803] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9208fd9-d326-4d44-9687-b1a677c4fbe3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1089.763045] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-969b58e3-59e5-4e68-add8-fe9840ca9419 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.769221] env[62385]: DEBUG oslo_vmware.api [None req-b9208fd9-d326-4d44-9687-b1a677c4fbe3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1089.769221] env[62385]: value = "task-1206485" [ 1089.769221] env[62385]: _type = "Task" [ 1089.769221] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.777464] env[62385]: DEBUG oslo_vmware.api [None req-b9208fd9-d326-4d44-9687-b1a677c4fbe3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206485, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.869780] env[62385]: DEBUG oslo_vmware.api [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206483, 'name': CloneVM_Task, 'duration_secs': 1.544871} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.870083] env[62385]: INFO nova.virt.vmwareapi.vmops [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Created linked-clone VM from snapshot [ 1089.870744] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aec6d239-e009-4be2-b364-d67e8f112ffb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.878409] env[62385]: DEBUG nova.virt.vmwareapi.images [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Uploading image 651910de-8449-4e93-bd88-46a95628369c {{(pid=62385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1089.895762] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: d646c9a3-0ed3-4e5b-9fc9-886a1a507f1b] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1089.906533] env[62385]: DEBUG oslo_vmware.rw_handles [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1089.906533] env[62385]: value = "vm-261314" [ 1089.906533] env[62385]: _type = "VirtualMachine" [ 1089.906533] env[62385]: }. {{(pid=62385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1089.906888] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-c8735901-e90a-4f7f-b5a1-685f2f414b78 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.914445] env[62385]: DEBUG oslo_vmware.rw_handles [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Lease: (returnval){ [ 1089.914445] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521642e3-6408-ac9e-9b58-ee5e02854ddd" [ 1089.914445] env[62385]: _type = "HttpNfcLease" [ 1089.914445] env[62385]: } obtained for exporting VM: (result){ [ 1089.914445] env[62385]: value = "vm-261314" [ 1089.914445] env[62385]: _type = "VirtualMachine" [ 1089.914445] env[62385]: }. {{(pid=62385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1089.914868] env[62385]: DEBUG oslo_vmware.api [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for the lease: (returnval){ [ 1089.914868] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521642e3-6408-ac9e-9b58-ee5e02854ddd" [ 1089.914868] env[62385]: _type = "HttpNfcLease" [ 1089.914868] env[62385]: } to be ready. {{(pid=62385) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1089.920920] env[62385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1089.920920] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521642e3-6408-ac9e-9b58-ee5e02854ddd" [ 1089.920920] env[62385]: _type = "HttpNfcLease" [ 1089.920920] env[62385]: } is initializing. {{(pid=62385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1089.942971] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8fc31933-dad3-45ab-8818-6c7ed1cffb50 tempest-AttachInterfacesTestJSON-1765133084 tempest-AttachInterfacesTestJSON-1765133084-project-member] Lock "57637d6e-39e2-405c-bbcf-64bf0ffbebd1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.146s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.953374] env[62385]: DEBUG oslo_vmware.api [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206484, 'name': ReconfigVM_Task, 'duration_secs': 0.271074} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.954020] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Reconfigured VM instance instance-00000070 to attach disk [datastore2] volume-6c3e125e-0b69-44e6-9e1a-3527fedf6e2a/volume-6c3e125e-0b69-44e6-9e1a-3527fedf6e2a.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1089.959134] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fffdc5e5-0424-4dfb-a0a1-ad1ac8575b54 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.973833] env[62385]: DEBUG oslo_vmware.api [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1089.973833] env[62385]: value = "task-1206487" [ 1089.973833] env[62385]: _type = "Task" [ 1089.973833] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.982495] env[62385]: DEBUG oslo_vmware.api [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206487, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.279602] env[62385]: DEBUG oslo_vmware.api [None req-b9208fd9-d326-4d44-9687-b1a677c4fbe3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206485, 'name': PowerOffVM_Task, 'duration_secs': 0.180347} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.279879] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9208fd9-d326-4d44-9687-b1a677c4fbe3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1090.280069] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-b9208fd9-d326-4d44-9687-b1a677c4fbe3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1090.280323] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dd3d4475-8516-4879-92d9-7d0ceec435a4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.348095] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-b9208fd9-d326-4d44-9687-b1a677c4fbe3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1090.348345] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-b9208fd9-d326-4d44-9687-b1a677c4fbe3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1090.348535] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9208fd9-d326-4d44-9687-b1a677c4fbe3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Deleting the datastore file [datastore1] 8e1d416c-2624-468c-94d7-6265e9f4178f {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1090.348842] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0b6a6499-99fd-4fd5-b3a3-3f6836ba6ad0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.355820] env[62385]: DEBUG oslo_vmware.api [None req-b9208fd9-d326-4d44-9687-b1a677c4fbe3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1090.355820] env[62385]: value = "task-1206489" [ 1090.355820] env[62385]: _type = "Task" [ 1090.355820] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.364272] env[62385]: DEBUG oslo_vmware.api [None req-b9208fd9-d326-4d44-9687-b1a677c4fbe3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206489, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.401258] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: fd7a77e2-6d76-4d91-bdef-e30333247aa9] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1090.422921] env[62385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1090.422921] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521642e3-6408-ac9e-9b58-ee5e02854ddd" [ 1090.422921] env[62385]: _type = "HttpNfcLease" [ 1090.422921] env[62385]: } is ready. {{(pid=62385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1090.423518] env[62385]: DEBUG oslo_vmware.rw_handles [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1090.423518] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521642e3-6408-ac9e-9b58-ee5e02854ddd" [ 1090.423518] env[62385]: _type = "HttpNfcLease" [ 1090.423518] env[62385]: }. {{(pid=62385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1090.423989] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d0f4277-adf7-4f71-b334-50762e68a798 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.431639] env[62385]: DEBUG oslo_vmware.rw_handles [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522b4d57-132c-e2e2-4550-31da646fd197/disk-0.vmdk from lease info. {{(pid=62385) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1090.431836] env[62385]: DEBUG oslo_vmware.rw_handles [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522b4d57-132c-e2e2-4550-31da646fd197/disk-0.vmdk for reading. {{(pid=62385) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1090.502462] env[62385]: DEBUG oslo_vmware.api [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206487, 'name': ReconfigVM_Task, 'duration_secs': 0.130204} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.502809] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261311', 'volume_id': '6c3e125e-0b69-44e6-9e1a-3527fedf6e2a', 'name': 'volume-6c3e125e-0b69-44e6-9e1a-3527fedf6e2a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c4f0482d-5b46-4c28-9c40-a59b6aee7dc3', 'attached_at': '', 'detached_at': '', 'volume_id': '6c3e125e-0b69-44e6-9e1a-3527fedf6e2a', 'serial': '6c3e125e-0b69-44e6-9e1a-3527fedf6e2a'} {{(pid=62385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1090.503685] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3dac2703-d8a0-4482-83ec-a6372b7c8e49 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.511231] env[62385]: DEBUG oslo_vmware.api [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1090.511231] env[62385]: value = "task-1206490" [ 1090.511231] env[62385]: _type = "Task" [ 1090.511231] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.520825] env[62385]: DEBUG oslo_vmware.api [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206490, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.523532] env[62385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-9ab86dc4-3b1a-4eed-a342-f2b509b36905 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.866983] env[62385]: DEBUG oslo_vmware.api [None req-b9208fd9-d326-4d44-9687-b1a677c4fbe3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206489, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.131672} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.867357] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9208fd9-d326-4d44-9687-b1a677c4fbe3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1090.867572] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-b9208fd9-d326-4d44-9687-b1a677c4fbe3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1090.867848] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-b9208fd9-d326-4d44-9687-b1a677c4fbe3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1090.868084] env[62385]: INFO nova.compute.manager [None req-b9208fd9-d326-4d44-9687-b1a677c4fbe3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1090.868518] env[62385]: DEBUG oslo.service.loopingcall [None req-b9208fd9-d326-4d44-9687-b1a677c4fbe3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1090.868772] env[62385]: DEBUG nova.compute.manager [-] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1090.869022] env[62385]: DEBUG nova.network.neutron [-] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1090.904875] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: b4970e1a-2b29-44e9-b79e-cda3c32dddcd] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1091.024472] env[62385]: DEBUG oslo_vmware.api [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206490, 'name': Rename_Task, 'duration_secs': 0.124651} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.030629] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1091.030629] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ee35367d-e4ce-455b-be9e-f0dd4fc6b695 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.036159] env[62385]: DEBUG oslo_vmware.api [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1091.036159] env[62385]: value = "task-1206491" [ 1091.036159] env[62385]: _type = "Task" [ 1091.036159] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.044862] env[62385]: DEBUG oslo_vmware.api [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206491, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.409749] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 0d823fcc-cbba-41f6-a47c-2f57d3e79948] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1091.512764] env[62385]: DEBUG nova.compute.manager [req-60ad586c-de69-48c9-b528-98c06292b174 req-0a5d839b-efaa-4bde-8d39-1c68a7af85c2 service nova] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Received event network-vif-deleted-deafdeb9-f59c-48f4-873b-6e4b27d4a688 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1091.512985] env[62385]: INFO nova.compute.manager [req-60ad586c-de69-48c9-b528-98c06292b174 req-0a5d839b-efaa-4bde-8d39-1c68a7af85c2 service nova] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Neutron deleted interface deafdeb9-f59c-48f4-873b-6e4b27d4a688; detaching it from the instance and deleting it from the info cache [ 1091.513627] env[62385]: DEBUG nova.network.neutron [req-60ad586c-de69-48c9-b528-98c06292b174 req-0a5d839b-efaa-4bde-8d39-1c68a7af85c2 service nova] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1091.553999] env[62385]: DEBUG oslo_vmware.api [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206491, 'name': PowerOnVM_Task, 'duration_secs': 0.482933} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.554447] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1091.554690] env[62385]: INFO nova.compute.manager [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Took 7.52 seconds to spawn the instance on the hypervisor. [ 1091.554875] env[62385]: DEBUG nova.compute.manager [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1091.557170] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cc1e365-6a3f-4177-bf48-7c797bc494b5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.913402] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 8d8e3b1a-350f-4bfd-9b41-4779cf1d12b7] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1091.975471] env[62385]: DEBUG nova.network.neutron [-] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1092.020027] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9b1bec9f-ea65-4cbf-af15-1a8074734ca7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.029621] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8bef6e6-1bc0-46dd-8857-6afaa800407b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.061834] env[62385]: DEBUG nova.compute.manager [req-60ad586c-de69-48c9-b528-98c06292b174 req-0a5d839b-efaa-4bde-8d39-1c68a7af85c2 service nova] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Detach interface failed, port_id=deafdeb9-f59c-48f4-873b-6e4b27d4a688, reason: Instance 8e1d416c-2624-468c-94d7-6265e9f4178f could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1092.088147] env[62385]: INFO nova.compute.manager [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Took 14.09 seconds to build instance. [ 1092.327996] env[62385]: DEBUG nova.compute.manager [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Stashing vm_state: active {{(pid=62385) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1092.419660] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: f77dc5ab-7e63-4a1b-9c87-e474a7e28c93] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1092.480067] env[62385]: INFO nova.compute.manager [-] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Took 1.61 seconds to deallocate network for instance. [ 1092.591458] env[62385]: DEBUG oslo_concurrency.lockutils [None req-46470d27-7e2d-45af-aa3f-ace9bd03ed40 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "c4f0482d-5b46-4c28-9c40-a59b6aee7dc3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.607s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.856943] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1092.857267] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1092.926377] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 65bb1d14-84b2-4e75-acdc-dc674a035101] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1092.987140] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b9208fd9-d326-4d44-9687-b1a677c4fbe3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.196336] env[62385]: DEBUG nova.compute.manager [req-15c26c60-c3a9-4acc-a962-fab5626f2d83 req-1793fdfd-2ad4-403d-a2f7-b0f8cd1f9519 service nova] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Received event network-changed-ff03b686-6596-418e-b748-0a8a6799a2b5 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1093.196536] env[62385]: DEBUG nova.compute.manager [req-15c26c60-c3a9-4acc-a962-fab5626f2d83 req-1793fdfd-2ad4-403d-a2f7-b0f8cd1f9519 service nova] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Refreshing instance network info cache due to event network-changed-ff03b686-6596-418e-b748-0a8a6799a2b5. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1093.197195] env[62385]: DEBUG oslo_concurrency.lockutils [req-15c26c60-c3a9-4acc-a962-fab5626f2d83 req-1793fdfd-2ad4-403d-a2f7-b0f8cd1f9519 service nova] Acquiring lock "refresh_cache-9d6f098a-0b05-43ef-96b0-9eb99ad3538c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1093.197367] env[62385]: DEBUG oslo_concurrency.lockutils [req-15c26c60-c3a9-4acc-a962-fab5626f2d83 req-1793fdfd-2ad4-403d-a2f7-b0f8cd1f9519 service nova] Acquired lock "refresh_cache-9d6f098a-0b05-43ef-96b0-9eb99ad3538c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.197529] env[62385]: DEBUG nova.network.neutron [req-15c26c60-c3a9-4acc-a962-fab5626f2d83 req-1793fdfd-2ad4-403d-a2f7-b0f8cd1f9519 service nova] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Refreshing network info cache for port ff03b686-6596-418e-b748-0a8a6799a2b5 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1093.362676] env[62385]: INFO nova.compute.claims [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1093.430021] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: e73ecefc-fed3-4ccd-88d3-399a4b72bb4b] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1093.868922] env[62385]: INFO nova.compute.resource_tracker [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Updating resource usage from migration c894b6b6-f1f1-4d64-b8a8-2e573459e7f3 [ 1093.914823] env[62385]: DEBUG nova.network.neutron [req-15c26c60-c3a9-4acc-a962-fab5626f2d83 req-1793fdfd-2ad4-403d-a2f7-b0f8cd1f9519 service nova] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Updated VIF entry in instance network info cache for port ff03b686-6596-418e-b748-0a8a6799a2b5. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1093.915280] env[62385]: DEBUG nova.network.neutron [req-15c26c60-c3a9-4acc-a962-fab5626f2d83 req-1793fdfd-2ad4-403d-a2f7-b0f8cd1f9519 service nova] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Updating instance_info_cache with network_info: [{"id": "ff03b686-6596-418e-b748-0a8a6799a2b5", "address": "fa:16:3e:91:05:f2", "network": {"id": "afae6b07-d08c-413f-aa73-b545db1bb766", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1654965885-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "93111f5ba17c47de9691f90b99e03aa8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff03b686-65", "ovs_interfaceid": "ff03b686-6596-418e-b748-0a8a6799a2b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.933052] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: a5c4afc6-38a6-4815-8ec4-cc01c24489bf] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1093.996833] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c07b32d0-9c46-43b9-aa3b-6a2c4140edd9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.004815] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad0cef46-8677-4c71-ad01-bd85a25fcf9b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.035778] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96a9da4d-b484-4d61-8613-f5577bd7bf3b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.043857] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2415f2e0-6d28-4525-9325-72cd80249022 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.058452] env[62385]: DEBUG nova.compute.provider_tree [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1094.420242] env[62385]: DEBUG oslo_concurrency.lockutils [req-15c26c60-c3a9-4acc-a962-fab5626f2d83 req-1793fdfd-2ad4-403d-a2f7-b0f8cd1f9519 service nova] Releasing lock "refresh_cache-9d6f098a-0b05-43ef-96b0-9eb99ad3538c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1094.439119] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 42ecf594-dff7-4af3-ac56-5bbc45b7a192] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1094.561681] env[62385]: DEBUG nova.scheduler.client.report [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1094.648260] env[62385]: DEBUG nova.compute.manager [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Stashing vm_state: active {{(pid=62385) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1094.943063] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 92057af7-28a3-4643-9cda-d3d868d0cfc9] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1095.068090] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.211s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.068358] env[62385]: INFO nova.compute.manager [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Migrating [ 1095.075694] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b9208fd9-d326-4d44-9687-b1a677c4fbe3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.089s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.075959] env[62385]: DEBUG nova.objects.instance [None req-b9208fd9-d326-4d44-9687-b1a677c4fbe3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lazy-loading 'resources' on Instance uuid 8e1d416c-2624-468c-94d7-6265e9f4178f {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1095.170661] env[62385]: DEBUG oslo_concurrency.lockutils [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.224614] env[62385]: DEBUG nova.compute.manager [req-2f799a8d-24c6-46c3-a888-f5b7070ae03e req-e538977d-5742-4907-a9a2-4dafa6cd7aee service nova] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Received event network-changed-07abbf5c-f2dc-4f07-a1a8-f81f59c6a2ae {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1095.224792] env[62385]: DEBUG nova.compute.manager [req-2f799a8d-24c6-46c3-a888-f5b7070ae03e req-e538977d-5742-4907-a9a2-4dafa6cd7aee service nova] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Refreshing instance network info cache due to event network-changed-07abbf5c-f2dc-4f07-a1a8-f81f59c6a2ae. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1095.225033] env[62385]: DEBUG oslo_concurrency.lockutils [req-2f799a8d-24c6-46c3-a888-f5b7070ae03e req-e538977d-5742-4907-a9a2-4dafa6cd7aee service nova] Acquiring lock "refresh_cache-c4f0482d-5b46-4c28-9c40-a59b6aee7dc3" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1095.225248] env[62385]: DEBUG oslo_concurrency.lockutils [req-2f799a8d-24c6-46c3-a888-f5b7070ae03e req-e538977d-5742-4907-a9a2-4dafa6cd7aee service nova] Acquired lock "refresh_cache-c4f0482d-5b46-4c28-9c40-a59b6aee7dc3" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.225411] env[62385]: DEBUG nova.network.neutron [req-2f799a8d-24c6-46c3-a888-f5b7070ae03e req-e538977d-5742-4907-a9a2-4dafa6cd7aee service nova] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Refreshing network info cache for port 07abbf5c-f2dc-4f07-a1a8-f81f59c6a2ae {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1095.447229] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 749e0976-2e2b-4764-865c-2e630f2edbd1] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1095.587220] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "refresh_cache-1289e63a-5a9d-4615-b6cf-bbf5f4ae0317" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1095.587415] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquired lock "refresh_cache-1289e63a-5a9d-4615-b6cf-bbf5f4ae0317" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.587762] env[62385]: DEBUG nova.network.neutron [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1095.696374] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-786a7cf2-fae2-45a3-a1db-dd88c29d2a63 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.705059] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99b8c4b4-b607-4ad0-8ab2-477ccd49f440 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.738316] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-764b47de-1575-494a-8277-6aa2fa4228db {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.745764] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ac2e5f8-3cc0-4711-bf51-75ec97ab6db8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.759112] env[62385]: DEBUG nova.compute.provider_tree [None req-b9208fd9-d326-4d44-9687-b1a677c4fbe3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1095.950570] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: e97ff020-61f3-4947-bb82-5c039ad17747] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1095.969091] env[62385]: DEBUG nova.network.neutron [req-2f799a8d-24c6-46c3-a888-f5b7070ae03e req-e538977d-5742-4907-a9a2-4dafa6cd7aee service nova] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Updated VIF entry in instance network info cache for port 07abbf5c-f2dc-4f07-a1a8-f81f59c6a2ae. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1095.969559] env[62385]: DEBUG nova.network.neutron [req-2f799a8d-24c6-46c3-a888-f5b7070ae03e req-e538977d-5742-4907-a9a2-4dafa6cd7aee service nova] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Updating instance_info_cache with network_info: [{"id": "07abbf5c-f2dc-4f07-a1a8-f81f59c6a2ae", "address": "fa:16:3e:a4:98:71", "network": {"id": "afae6b07-d08c-413f-aa73-b545db1bb766", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1654965885-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.163", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "93111f5ba17c47de9691f90b99e03aa8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap07abbf5c-f2", "ovs_interfaceid": "07abbf5c-f2dc-4f07-a1a8-f81f59c6a2ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1096.262610] env[62385]: DEBUG nova.scheduler.client.report [None req-b9208fd9-d326-4d44-9687-b1a677c4fbe3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1096.318800] env[62385]: DEBUG nova.network.neutron [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Updating instance_info_cache with network_info: [{"id": "c6ede9da-b209-4fae-b75a-c2f1ede4dcaf", "address": "fa:16:3e:a7:e1:c0", "network": {"id": "06017161-984c-4657-9639-4e6bdb6d4c09", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1072882119-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.208", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d45b5346eab44707b4d81eaaf12b0935", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39cd75b0-9ec7-48ed-b57f-34da0c573a60", "external-id": "nsx-vlan-transportzone-751", "segmentation_id": 751, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6ede9da-b2", "ovs_interfaceid": "c6ede9da-b209-4fae-b75a-c2f1ede4dcaf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1096.454023] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: caca5304-0ad7-427c-812e-de925de63f2c] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1096.472495] env[62385]: DEBUG oslo_concurrency.lockutils [req-2f799a8d-24c6-46c3-a888-f5b7070ae03e req-e538977d-5742-4907-a9a2-4dafa6cd7aee service nova] Releasing lock "refresh_cache-c4f0482d-5b46-4c28-9c40-a59b6aee7dc3" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1096.768102] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b9208fd9-d326-4d44-9687-b1a677c4fbe3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.692s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.771528] env[62385]: DEBUG oslo_concurrency.lockutils [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 1.601s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.785713] env[62385]: INFO nova.scheduler.client.report [None req-b9208fd9-d326-4d44-9687-b1a677c4fbe3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Deleted allocations for instance 8e1d416c-2624-468c-94d7-6265e9f4178f [ 1096.821987] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Releasing lock "refresh_cache-1289e63a-5a9d-4615-b6cf-bbf5f4ae0317" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1096.958031] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 8a099f6e-b5e6-4e3f-a8ab-abdae59d73ee] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1097.276506] env[62385]: INFO nova.compute.claims [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1097.294299] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b9208fd9-d326-4d44-9687-b1a677c4fbe3 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lock "8e1d416c-2624-468c-94d7-6265e9f4178f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.546s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.461416] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: f50ddafe-f5a9-422b-b0f4-46c1b111dfbe] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1097.783016] env[62385]: INFO nova.compute.resource_tracker [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Updating resource usage from migration 088b8d70-a0fe-47d3-a27c-c76757ffc5bd [ 1097.880067] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7c67129-d54b-4ee0-96e5-b4ef0487b23b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.888444] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77e1f2b4-2838-4cf3-8722-2f61b0ae097b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.917582] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb803e03-3657-461b-a68d-2b1615a8871f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.924725] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-158bcaef-aded-42e2-a7e6-9308d4975dd3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.937882] env[62385]: DEBUG nova.compute.provider_tree [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1097.964862] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: f1d33401-35f5-4d79-abb4-26dc6faa784e] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1098.336970] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f45a5695-2154-48fc-989b-a7207114918b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.355865] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Updating instance '1289e63a-5a9d-4615-b6cf-bbf5f4ae0317' progress to 0 {{(pid=62385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1098.382490] env[62385]: DEBUG oslo_vmware.rw_handles [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522b4d57-132c-e2e2-4550-31da646fd197/disk-0.vmdk. {{(pid=62385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1098.383508] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-726ec414-efb3-4539-b8f1-b7a3ac727dc5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.389622] env[62385]: DEBUG oslo_vmware.rw_handles [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522b4d57-132c-e2e2-4550-31da646fd197/disk-0.vmdk is in state: ready. {{(pid=62385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1098.389795] env[62385]: ERROR oslo_vmware.rw_handles [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522b4d57-132c-e2e2-4550-31da646fd197/disk-0.vmdk due to incomplete transfer. [ 1098.390032] env[62385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-3e2bfde2-5902-481f-872e-c9f703b2f1a0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.397393] env[62385]: DEBUG oslo_vmware.rw_handles [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522b4d57-132c-e2e2-4550-31da646fd197/disk-0.vmdk. {{(pid=62385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1098.397599] env[62385]: DEBUG nova.virt.vmwareapi.images [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Uploaded image 651910de-8449-4e93-bd88-46a95628369c to the Glance image server {{(pid=62385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1098.399911] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Destroying the VM {{(pid=62385) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1098.400839] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-ebff9e20-6fef-4e0a-9d7e-520ddd032bbc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.408012] env[62385]: DEBUG oslo_vmware.api [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for the task: (returnval){ [ 1098.408012] env[62385]: value = "task-1206492" [ 1098.408012] env[62385]: _type = "Task" [ 1098.408012] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.415687] env[62385]: DEBUG oslo_vmware.api [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206492, 'name': Destroy_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.440586] env[62385]: DEBUG nova.scheduler.client.report [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1098.467912] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 33387505-c576-488b-8c9c-b064fe81a7d7] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1098.862653] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1098.862934] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6bb8aa08-0202-4143-a300-9ad857f72fc9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.869948] env[62385]: DEBUG oslo_vmware.api [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1098.869948] env[62385]: value = "task-1206493" [ 1098.869948] env[62385]: _type = "Task" [ 1098.869948] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.878060] env[62385]: DEBUG oslo_vmware.api [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206493, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.918365] env[62385]: DEBUG oslo_vmware.api [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206492, 'name': Destroy_Task, 'duration_secs': 0.313492} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.919448] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Destroyed the VM [ 1098.919448] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Deleting Snapshot of the VM instance {{(pid=62385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1098.919448] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-43e416f2-5be5-4276-8a33-f1741aa28b38 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.925970] env[62385]: DEBUG oslo_vmware.api [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for the task: (returnval){ [ 1098.925970] env[62385]: value = "task-1206494" [ 1098.925970] env[62385]: _type = "Task" [ 1098.925970] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.934738] env[62385]: DEBUG oslo_vmware.api [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206494, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.945412] env[62385]: DEBUG oslo_concurrency.lockutils [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.174s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.945664] env[62385]: INFO nova.compute.manager [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Migrating [ 1098.971406] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 3e184125-28af-469b-83cc-4ab8859e0e2f] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1099.216882] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquiring lock "4fe375be-ba60-4601-bd89-610e12742df3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1099.217138] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lock "4fe375be-ba60-4601-bd89-610e12742df3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1099.380298] env[62385]: DEBUG oslo_vmware.api [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206493, 'name': PowerOffVM_Task, 'duration_secs': 0.174394} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.380522] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1099.380720] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Updating instance '1289e63a-5a9d-4615-b6cf-bbf5f4ae0317' progress to 17 {{(pid=62385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1099.436775] env[62385]: DEBUG oslo_vmware.api [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206494, 'name': RemoveSnapshot_Task, 'duration_secs': 0.389073} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.437053] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Deleted Snapshot of the VM instance {{(pid=62385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1099.437343] env[62385]: DEBUG nova.compute.manager [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1099.438103] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08058337-b329-442a-83e4-64b454b2f171 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.460412] env[62385]: DEBUG oslo_concurrency.lockutils [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "refresh_cache-c4f0482d-5b46-4c28-9c40-a59b6aee7dc3" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1099.460636] env[62385]: DEBUG oslo_concurrency.lockutils [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquired lock "refresh_cache-c4f0482d-5b46-4c28-9c40-a59b6aee7dc3" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1099.460850] env[62385]: DEBUG nova.network.neutron [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1099.474582] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 4a4b254c-46a3-4db6-9c0f-f476ae8f2b69] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1099.720222] env[62385]: DEBUG nova.compute.manager [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1099.887712] env[62385]: DEBUG nova.virt.hardware [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:16:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1099.888009] env[62385]: DEBUG nova.virt.hardware [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1099.888289] env[62385]: DEBUG nova.virt.hardware [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1099.888510] env[62385]: DEBUG nova.virt.hardware [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1099.888669] env[62385]: DEBUG nova.virt.hardware [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1099.888827] env[62385]: DEBUG nova.virt.hardware [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1099.889052] env[62385]: DEBUG nova.virt.hardware [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1099.889233] env[62385]: DEBUG nova.virt.hardware [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1099.889407] env[62385]: DEBUG nova.virt.hardware [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1099.889574] env[62385]: DEBUG nova.virt.hardware [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1099.889753] env[62385]: DEBUG nova.virt.hardware [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1099.894856] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d04fc1da-15a4-4232-8298-e11fd14972a1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.910964] env[62385]: DEBUG oslo_vmware.api [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1099.910964] env[62385]: value = "task-1206495" [ 1099.910964] env[62385]: _type = "Task" [ 1099.910964] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.918546] env[62385]: DEBUG oslo_vmware.api [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206495, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.949992] env[62385]: INFO nova.compute.manager [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Shelve offloading [ 1099.951882] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1099.952160] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0ee9227f-127b-436d-ae79-c229f3675b0d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.958366] env[62385]: DEBUG oslo_vmware.api [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for the task: (returnval){ [ 1099.958366] env[62385]: value = "task-1206496" [ 1099.958366] env[62385]: _type = "Task" [ 1099.958366] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.967814] env[62385]: DEBUG oslo_vmware.api [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206496, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.978582] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 1ddb4d6d-3872-49db-bb40-e21721241e89] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1100.169953] env[62385]: DEBUG nova.network.neutron [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Updating instance_info_cache with network_info: [{"id": "07abbf5c-f2dc-4f07-a1a8-f81f59c6a2ae", "address": "fa:16:3e:a4:98:71", "network": {"id": "afae6b07-d08c-413f-aa73-b545db1bb766", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1654965885-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.163", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "93111f5ba17c47de9691f90b99e03aa8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap07abbf5c-f2", "ovs_interfaceid": "07abbf5c-f2dc-4f07-a1a8-f81f59c6a2ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1100.242589] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1100.242939] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1100.244593] env[62385]: INFO nova.compute.claims [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1100.420723] env[62385]: DEBUG oslo_vmware.api [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206495, 'name': ReconfigVM_Task, 'duration_secs': 0.489494} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.421032] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Updating instance '1289e63a-5a9d-4615-b6cf-bbf5f4ae0317' progress to 33 {{(pid=62385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1100.469459] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] VM already powered off {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1100.469459] env[62385]: DEBUG nova.compute.manager [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1100.470155] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45ff1013-5586-490d-9d4f-c6bd891df0b6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.476377] env[62385]: DEBUG oslo_concurrency.lockutils [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Acquiring lock "refresh_cache-c49d4366-2ea0-44f2-8463-daf62bd1c40c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1100.476538] env[62385]: DEBUG oslo_concurrency.lockutils [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Acquired lock "refresh_cache-c49d4366-2ea0-44f2-8463-daf62bd1c40c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1100.476704] env[62385]: DEBUG nova.network.neutron [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1100.481840] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1100.482604] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Cleaning up deleted instances with incomplete migration {{(pid=62385) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1100.674071] env[62385]: DEBUG oslo_concurrency.lockutils [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Releasing lock "refresh_cache-c4f0482d-5b46-4c28-9c40-a59b6aee7dc3" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1100.926966] env[62385]: DEBUG nova.virt.hardware [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1100.927233] env[62385]: DEBUG nova.virt.hardware [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1100.927400] env[62385]: DEBUG nova.virt.hardware [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1100.927590] env[62385]: DEBUG nova.virt.hardware [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1100.927739] env[62385]: DEBUG nova.virt.hardware [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1100.927890] env[62385]: DEBUG nova.virt.hardware [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1100.928111] env[62385]: DEBUG nova.virt.hardware [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1100.928280] env[62385]: DEBUG nova.virt.hardware [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1100.928455] env[62385]: DEBUG nova.virt.hardware [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1100.928628] env[62385]: DEBUG nova.virt.hardware [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1100.928798] env[62385]: DEBUG nova.virt.hardware [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1100.934217] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Reconfiguring VM instance instance-0000006c to detach disk 2000 {{(pid=62385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1100.934505] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7e127b07-0321-4201-9939-654b6a7d3f75 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.952427] env[62385]: DEBUG oslo_vmware.api [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1100.952427] env[62385]: value = "task-1206497" [ 1100.952427] env[62385]: _type = "Task" [ 1100.952427] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.959951] env[62385]: DEBUG oslo_vmware.api [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206497, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.986316] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1101.166491] env[62385]: DEBUG nova.network.neutron [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Updating instance_info_cache with network_info: [{"id": "b19d5597-14de-4b08-b628-3bc8e539f77a", "address": "fa:16:3e:83:b5:c7", "network": {"id": "a561a523-c9d8-4169-b449-5c0bdef7b53b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-751953866-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d716d19f79945db98f684df0cf4b302", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb19d5597-14", "ovs_interfaceid": "b19d5597-14de-4b08-b628-3bc8e539f77a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1101.344617] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11263dc6-253a-42dc-810b-dc705efc546e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.352333] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57f51e16-5dea-4a7d-b71c-f794807e0915 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.382169] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33eed33d-713e-4ae3-aebf-b8fbc4fb0a15 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.389012] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-359536de-7640-4a7f-b225-effa0ed671dc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.401165] env[62385]: DEBUG nova.compute.provider_tree [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1101.461825] env[62385]: DEBUG oslo_vmware.api [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206497, 'name': ReconfigVM_Task, 'duration_secs': 0.163367} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.462113] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Reconfigured VM instance instance-0000006c to detach disk 2000 {{(pid=62385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1101.462882] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0b86824-709c-405d-afe7-b0c8048ed50b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.485087] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317/1289e63a-5a9d-4615-b6cf-bbf5f4ae0317.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1101.485326] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3cb077d8-361d-4b2f-ae39-8556627583a9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.503317] env[62385]: DEBUG oslo_vmware.api [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1101.503317] env[62385]: value = "task-1206498" [ 1101.503317] env[62385]: _type = "Task" [ 1101.503317] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.510170] env[62385]: DEBUG oslo_vmware.api [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206498, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.669737] env[62385]: DEBUG oslo_concurrency.lockutils [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Releasing lock "refresh_cache-c49d4366-2ea0-44f2-8463-daf62bd1c40c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1101.904530] env[62385]: DEBUG nova.scheduler.client.report [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1101.916453] env[62385]: DEBUG nova.compute.manager [req-206aa932-1d7d-4656-816e-2f153e4a6c65 req-6f08260f-ec97-4439-b1c4-9676d5f44e3d service nova] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Received event network-vif-unplugged-b19d5597-14de-4b08-b628-3bc8e539f77a {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1101.917403] env[62385]: DEBUG oslo_concurrency.lockutils [req-206aa932-1d7d-4656-816e-2f153e4a6c65 req-6f08260f-ec97-4439-b1c4-9676d5f44e3d service nova] Acquiring lock "c49d4366-2ea0-44f2-8463-daf62bd1c40c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.917403] env[62385]: DEBUG oslo_concurrency.lockutils [req-206aa932-1d7d-4656-816e-2f153e4a6c65 req-6f08260f-ec97-4439-b1c4-9676d5f44e3d service nova] Lock "c49d4366-2ea0-44f2-8463-daf62bd1c40c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.917403] env[62385]: DEBUG oslo_concurrency.lockutils [req-206aa932-1d7d-4656-816e-2f153e4a6c65 req-6f08260f-ec97-4439-b1c4-9676d5f44e3d service nova] Lock "c49d4366-2ea0-44f2-8463-daf62bd1c40c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1101.917403] env[62385]: DEBUG nova.compute.manager [req-206aa932-1d7d-4656-816e-2f153e4a6c65 req-6f08260f-ec97-4439-b1c4-9676d5f44e3d service nova] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] No waiting events found dispatching network-vif-unplugged-b19d5597-14de-4b08-b628-3bc8e539f77a {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1101.917403] env[62385]: WARNING nova.compute.manager [req-206aa932-1d7d-4656-816e-2f153e4a6c65 req-6f08260f-ec97-4439-b1c4-9676d5f44e3d service nova] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Received unexpected event network-vif-unplugged-b19d5597-14de-4b08-b628-3bc8e539f77a for instance with vm_state shelved and task_state shelving_offloading. [ 1101.943477] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1101.944383] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67694536-ed9e-499f-ae09-186c0647026d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.952088] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1101.952340] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-008c2755-5910-49bb-8298-cc90740494ec {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.012877] env[62385]: DEBUG oslo_vmware.api [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206498, 'name': ReconfigVM_Task, 'duration_secs': 0.240258} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.013945] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Reconfigured VM instance instance-0000006c to attach disk [datastore1] 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317/1289e63a-5a9d-4615-b6cf-bbf5f4ae0317.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1102.014238] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Updating instance '1289e63a-5a9d-4615-b6cf-bbf5f4ae0317' progress to 50 {{(pid=62385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1102.017788] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1102.017886] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1102.018079] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Deleting the datastore file [datastore1] c49d4366-2ea0-44f2-8463-daf62bd1c40c {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1102.018585] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7558a199-1de6-4542-80c2-4912d6f3b9b7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.023904] env[62385]: DEBUG oslo_vmware.api [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for the task: (returnval){ [ 1102.023904] env[62385]: value = "task-1206500" [ 1102.023904] env[62385]: _type = "Task" [ 1102.023904] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.031420] env[62385]: DEBUG oslo_vmware.api [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206500, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.192701] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13d51fc6-b23f-46f4-9867-3996eefe683c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.210645] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Updating instance 'c4f0482d-5b46-4c28-9c40-a59b6aee7dc3' progress to 0 {{(pid=62385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1102.409246] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.166s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.409822] env[62385]: DEBUG nova.compute.manager [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1102.521103] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf5fc20e-9a0b-443a-9eac-c9759fb15e08 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.542378] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c86c9c9a-55f0-4fbc-8ec5-b066cad28bb9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.547235] env[62385]: DEBUG oslo_vmware.api [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206500, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.120961} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.547831] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1102.548029] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1102.548231] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1102.562694] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Updating instance '1289e63a-5a9d-4615-b6cf-bbf5f4ae0317' progress to 67 {{(pid=62385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1102.580764] env[62385]: INFO nova.scheduler.client.report [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Deleted allocations for instance c49d4366-2ea0-44f2-8463-daf62bd1c40c [ 1102.716165] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1102.716480] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-71f2d894-e88f-411c-8278-2b484d0026ac {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.725290] env[62385]: DEBUG oslo_vmware.api [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1102.725290] env[62385]: value = "task-1206501" [ 1102.725290] env[62385]: _type = "Task" [ 1102.725290] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.733780] env[62385]: DEBUG oslo_vmware.api [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206501, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.914629] env[62385]: DEBUG nova.compute.utils [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1102.916149] env[62385]: DEBUG nova.compute.manager [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1102.916335] env[62385]: DEBUG nova.network.neutron [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1102.956557] env[62385]: DEBUG nova.policy [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e56d4575033b4338acd01cdfa29f9693', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '09c2c5e938bf46fd9353e47aecc72478', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 1103.086095] env[62385]: DEBUG oslo_concurrency.lockutils [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1103.086397] env[62385]: DEBUG oslo_concurrency.lockutils [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1103.086612] env[62385]: DEBUG nova.objects.instance [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Lazy-loading 'resources' on Instance uuid c49d4366-2ea0-44f2-8463-daf62bd1c40c {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1103.102367] env[62385]: DEBUG nova.network.neutron [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Port c6ede9da-b209-4fae-b75a-c2f1ede4dcaf binding to destination host cpu-1 is already ACTIVE {{(pid=62385) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1103.237295] env[62385]: DEBUG oslo_vmware.api [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206501, 'name': PowerOffVM_Task, 'duration_secs': 0.202264} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.237595] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1103.237840] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Updating instance 'c4f0482d-5b46-4c28-9c40-a59b6aee7dc3' progress to 17 {{(pid=62385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1103.242209] env[62385]: DEBUG nova.network.neutron [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Successfully created port: 1e16fa5f-3cab-4323-8637-7ff6dffce79b {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1103.419409] env[62385]: DEBUG nova.compute.manager [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1103.588725] env[62385]: DEBUG nova.objects.instance [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Lazy-loading 'numa_topology' on Instance uuid c49d4366-2ea0-44f2-8463-daf62bd1c40c {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1103.747136] env[62385]: DEBUG nova.virt.hardware [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:16:00Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1103.747355] env[62385]: DEBUG nova.virt.hardware [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1103.747499] env[62385]: DEBUG nova.virt.hardware [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1103.747707] env[62385]: DEBUG nova.virt.hardware [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1103.747889] env[62385]: DEBUG nova.virt.hardware [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1103.748082] env[62385]: DEBUG nova.virt.hardware [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1103.748304] env[62385]: DEBUG nova.virt.hardware [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1103.748474] env[62385]: DEBUG nova.virt.hardware [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1103.748646] env[62385]: DEBUG nova.virt.hardware [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1103.748818] env[62385]: DEBUG nova.virt.hardware [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1103.748995] env[62385]: DEBUG nova.virt.hardware [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1103.754162] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2e9d64d1-3565-4d21-8724-af663f2dacb9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.769890] env[62385]: DEBUG oslo_vmware.api [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1103.769890] env[62385]: value = "task-1206502" [ 1103.769890] env[62385]: _type = "Task" [ 1103.769890] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.778120] env[62385]: DEBUG oslo_vmware.api [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206502, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.091450] env[62385]: DEBUG nova.objects.base [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=62385) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1104.129230] env[62385]: DEBUG nova.compute.manager [req-19fedac8-138d-4894-a93d-1841fe07e161 req-21b76995-faf4-491c-82bd-948ff3b424bc service nova] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Received event network-changed-b19d5597-14de-4b08-b628-3bc8e539f77a {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1104.129665] env[62385]: DEBUG nova.compute.manager [req-19fedac8-138d-4894-a93d-1841fe07e161 req-21b76995-faf4-491c-82bd-948ff3b424bc service nova] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Refreshing instance network info cache due to event network-changed-b19d5597-14de-4b08-b628-3bc8e539f77a. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1104.129817] env[62385]: DEBUG oslo_concurrency.lockutils [req-19fedac8-138d-4894-a93d-1841fe07e161 req-21b76995-faf4-491c-82bd-948ff3b424bc service nova] Acquiring lock "refresh_cache-c49d4366-2ea0-44f2-8463-daf62bd1c40c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1104.130169] env[62385]: DEBUG oslo_concurrency.lockutils [req-19fedac8-138d-4894-a93d-1841fe07e161 req-21b76995-faf4-491c-82bd-948ff3b424bc service nova] Acquired lock "refresh_cache-c49d4366-2ea0-44f2-8463-daf62bd1c40c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.130252] env[62385]: DEBUG nova.network.neutron [req-19fedac8-138d-4894-a93d-1841fe07e161 req-21b76995-faf4-491c-82bd-948ff3b424bc service nova] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Refreshing network info cache for port b19d5597-14de-4b08-b628-3bc8e539f77a {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1104.132483] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "1289e63a-5a9d-4615-b6cf-bbf5f4ae0317-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.132767] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "1289e63a-5a9d-4615-b6cf-bbf5f4ae0317-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1104.132995] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "1289e63a-5a9d-4615-b6cf-bbf5f4ae0317-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.208577] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-394b6a0c-a2bc-4d58-b742-13157633a8cb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.216379] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6826fd46-e93a-4d4c-bc78-a197e9bd4032 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.245999] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7fd6ede-89ac-4d8e-9988-341e39486842 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.253624] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bb8eb66-2cd1-41bf-a2b9-5b51f31ace44 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.268862] env[62385]: DEBUG nova.compute.provider_tree [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1104.278790] env[62385]: DEBUG oslo_vmware.api [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206502, 'name': ReconfigVM_Task, 'duration_secs': 0.180239} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.279091] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Updating instance 'c4f0482d-5b46-4c28-9c40-a59b6aee7dc3' progress to 33 {{(pid=62385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1104.393085] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Acquiring lock "c49d4366-2ea0-44f2-8463-daf62bd1c40c" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.429441] env[62385]: DEBUG nova.compute.manager [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1104.455845] env[62385]: DEBUG nova.virt.hardware [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1104.456113] env[62385]: DEBUG nova.virt.hardware [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1104.456283] env[62385]: DEBUG nova.virt.hardware [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1104.456471] env[62385]: DEBUG nova.virt.hardware [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1104.456622] env[62385]: DEBUG nova.virt.hardware [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1104.456776] env[62385]: DEBUG nova.virt.hardware [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1104.456986] env[62385]: DEBUG nova.virt.hardware [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1104.457163] env[62385]: DEBUG nova.virt.hardware [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1104.457336] env[62385]: DEBUG nova.virt.hardware [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1104.457503] env[62385]: DEBUG nova.virt.hardware [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1104.457677] env[62385]: DEBUG nova.virt.hardware [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1104.458546] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cb29d39-f2e2-4fc4-ab0d-3a8d072192c4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.466247] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c94f4f7a-bc03-4242-9a92-57927e31e00e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.737277] env[62385]: DEBUG nova.network.neutron [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Successfully updated port: 1e16fa5f-3cab-4323-8637-7ff6dffce79b {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1104.771951] env[62385]: DEBUG nova.scheduler.client.report [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1104.784325] env[62385]: DEBUG nova.virt.hardware [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1104.784558] env[62385]: DEBUG nova.virt.hardware [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1104.784721] env[62385]: DEBUG nova.virt.hardware [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1104.784908] env[62385]: DEBUG nova.virt.hardware [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1104.785075] env[62385]: DEBUG nova.virt.hardware [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1104.785233] env[62385]: DEBUG nova.virt.hardware [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1104.785440] env[62385]: DEBUG nova.virt.hardware [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1104.785635] env[62385]: DEBUG nova.virt.hardware [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1104.785774] env[62385]: DEBUG nova.virt.hardware [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1104.785939] env[62385]: DEBUG nova.virt.hardware [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1104.786137] env[62385]: DEBUG nova.virt.hardware [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1104.826794] env[62385]: DEBUG nova.network.neutron [req-19fedac8-138d-4894-a93d-1841fe07e161 req-21b76995-faf4-491c-82bd-948ff3b424bc service nova] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Updated VIF entry in instance network info cache for port b19d5597-14de-4b08-b628-3bc8e539f77a. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1104.827140] env[62385]: DEBUG nova.network.neutron [req-19fedac8-138d-4894-a93d-1841fe07e161 req-21b76995-faf4-491c-82bd-948ff3b424bc service nova] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Updating instance_info_cache with network_info: [{"id": "b19d5597-14de-4b08-b628-3bc8e539f77a", "address": "fa:16:3e:83:b5:c7", "network": {"id": "a561a523-c9d8-4169-b449-5c0bdef7b53b", "bridge": null, "label": "tempest-ServersNegativeTestJSON-751953866-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d716d19f79945db98f684df0cf4b302", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapb19d5597-14", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1105.195207] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "refresh_cache-1289e63a-5a9d-4615-b6cf-bbf5f4ae0317" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1105.195478] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquired lock "refresh_cache-1289e63a-5a9d-4615-b6cf-bbf5f4ae0317" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1105.195663] env[62385]: DEBUG nova.network.neutron [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1105.240115] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquiring lock "refresh_cache-4fe375be-ba60-4601-bd89-610e12742df3" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1105.240399] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquired lock "refresh_cache-4fe375be-ba60-4601-bd89-610e12742df3" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1105.240581] env[62385]: DEBUG nova.network.neutron [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1105.276469] env[62385]: DEBUG oslo_concurrency.lockutils [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.190s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.294545] env[62385]: ERROR nova.compute.manager [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Setting instance vm_state to ERROR: AttributeError: 'NoneType' object has no attribute 'key' [ 1105.294545] env[62385]: ERROR nova.compute.manager [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Traceback (most recent call last): [ 1105.294545] env[62385]: ERROR nova.compute.manager [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] File "/opt/stack/nova/nova/compute/manager.py", line 10865, in _error_out_instance_on_exception [ 1105.294545] env[62385]: ERROR nova.compute.manager [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] yield [ 1105.294545] env[62385]: ERROR nova.compute.manager [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] File "/opt/stack/nova/nova/compute/manager.py", line 6105, in _resize_instance [ 1105.294545] env[62385]: ERROR nova.compute.manager [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] disk_info = self.driver.migrate_disk_and_power_off( [ 1105.294545] env[62385]: ERROR nova.compute.manager [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 264, in migrate_disk_and_power_off [ 1105.294545] env[62385]: ERROR nova.compute.manager [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] return self._vmops.migrate_disk_and_power_off(context, instance, [ 1105.294545] env[62385]: ERROR nova.compute.manager [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1467, in migrate_disk_and_power_off [ 1105.294545] env[62385]: ERROR nova.compute.manager [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] self._resize_disk(instance, vm_ref, vmdk, flavor) [ 1105.294545] env[62385]: ERROR nova.compute.manager [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1398, in _resize_disk [ 1105.294545] env[62385]: ERROR nova.compute.manager [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] self._volumeops.detach_disk_from_vm(vm_ref, instance, vmdk.device) [ 1105.294545] env[62385]: ERROR nova.compute.manager [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 121, in detach_disk_from_vm [ 1105.294545] env[62385]: ERROR nova.compute.manager [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] disk_key = device.key [ 1105.294545] env[62385]: ERROR nova.compute.manager [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] AttributeError: 'NoneType' object has no attribute 'key' [ 1105.294545] env[62385]: ERROR nova.compute.manager [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] [ 1105.329827] env[62385]: DEBUG oslo_concurrency.lockutils [req-19fedac8-138d-4894-a93d-1841fe07e161 req-21b76995-faf4-491c-82bd-948ff3b424bc service nova] Releasing lock "refresh_cache-c49d4366-2ea0-44f2-8463-daf62bd1c40c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1105.772859] env[62385]: DEBUG nova.network.neutron [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1105.786495] env[62385]: DEBUG oslo_concurrency.lockutils [None req-73c6dbf3-3a82-4acd-bde3-91ce29b79ee4 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Lock "c49d4366-2ea0-44f2-8463-daf62bd1c40c" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 20.557s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.787435] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Lock "c49d4366-2ea0-44f2-8463-daf62bd1c40c" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.395s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.787826] env[62385]: INFO nova.compute.manager [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Unshelving [ 1105.816375] env[62385]: INFO nova.compute.manager [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Swapping old allocation on dict_keys(['8c6086eb-4f96-48ed-af58-d7d7f36e7c5a']) held by migration 088b8d70-a0fe-47d3-a27c-c76757ffc5bd for instance [ 1105.841472] env[62385]: DEBUG nova.scheduler.client.report [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Overwriting current allocation {'allocations': {'8c6086eb-4f96-48ed-af58-d7d7f36e7c5a': {'resources': {'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 147}}, 'project_id': '93111f5ba17c47de9691f90b99e03aa8', 'user_id': 'b314c392b9ca405480c27c25364fda81', 'consumer_generation': 1} on consumer c4f0482d-5b46-4c28-9c40-a59b6aee7dc3 {{(pid=62385) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1105.908855] env[62385]: DEBUG nova.network.neutron [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Updating instance_info_cache with network_info: [{"id": "1e16fa5f-3cab-4323-8637-7ff6dffce79b", "address": "fa:16:3e:6c:0c:dd", "network": {"id": "9c4d6464-cacc-44f8-b7e5-0699cd312c53", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1889153615-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09c2c5e938bf46fd9353e47aecc72478", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4d3f69a-b086-4c3b-b976-5a848b63dfc4", "external-id": "nsx-vlan-transportzone-627", "segmentation_id": 627, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e16fa5f-3c", "ovs_interfaceid": "1e16fa5f-3cab-4323-8637-7ff6dffce79b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1105.910469] env[62385]: DEBUG nova.network.neutron [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Updating instance_info_cache with network_info: [{"id": "c6ede9da-b209-4fae-b75a-c2f1ede4dcaf", "address": "fa:16:3e:a7:e1:c0", "network": {"id": "06017161-984c-4657-9639-4e6bdb6d4c09", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1072882119-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.208", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d45b5346eab44707b4d81eaaf12b0935", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39cd75b0-9ec7-48ed-b57f-34da0c573a60", "external-id": "nsx-vlan-transportzone-751", "segmentation_id": 751, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6ede9da-b2", "ovs_interfaceid": "c6ede9da-b209-4fae-b75a-c2f1ede4dcaf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1106.149224] env[62385]: DEBUG nova.compute.manager [req-f82e4972-30f2-4dd4-82ca-6a2444fdeb86 req-a6318424-0f46-46c6-bb13-eb9fb05ff816 service nova] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Received event network-vif-plugged-1e16fa5f-3cab-4323-8637-7ff6dffce79b {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1106.149224] env[62385]: DEBUG oslo_concurrency.lockutils [req-f82e4972-30f2-4dd4-82ca-6a2444fdeb86 req-a6318424-0f46-46c6-bb13-eb9fb05ff816 service nova] Acquiring lock "4fe375be-ba60-4601-bd89-610e12742df3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1106.149224] env[62385]: DEBUG oslo_concurrency.lockutils [req-f82e4972-30f2-4dd4-82ca-6a2444fdeb86 req-a6318424-0f46-46c6-bb13-eb9fb05ff816 service nova] Lock "4fe375be-ba60-4601-bd89-610e12742df3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1106.149347] env[62385]: DEBUG oslo_concurrency.lockutils [req-f82e4972-30f2-4dd4-82ca-6a2444fdeb86 req-a6318424-0f46-46c6-bb13-eb9fb05ff816 service nova] Lock "4fe375be-ba60-4601-bd89-610e12742df3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1106.149506] env[62385]: DEBUG nova.compute.manager [req-f82e4972-30f2-4dd4-82ca-6a2444fdeb86 req-a6318424-0f46-46c6-bb13-eb9fb05ff816 service nova] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] No waiting events found dispatching network-vif-plugged-1e16fa5f-3cab-4323-8637-7ff6dffce79b {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1106.149678] env[62385]: WARNING nova.compute.manager [req-f82e4972-30f2-4dd4-82ca-6a2444fdeb86 req-a6318424-0f46-46c6-bb13-eb9fb05ff816 service nova] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Received unexpected event network-vif-plugged-1e16fa5f-3cab-4323-8637-7ff6dffce79b for instance with vm_state building and task_state spawning. [ 1106.149843] env[62385]: DEBUG nova.compute.manager [req-f82e4972-30f2-4dd4-82ca-6a2444fdeb86 req-a6318424-0f46-46c6-bb13-eb9fb05ff816 service nova] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Received event network-changed-1e16fa5f-3cab-4323-8637-7ff6dffce79b {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1106.149998] env[62385]: DEBUG nova.compute.manager [req-f82e4972-30f2-4dd4-82ca-6a2444fdeb86 req-a6318424-0f46-46c6-bb13-eb9fb05ff816 service nova] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Refreshing instance network info cache due to event network-changed-1e16fa5f-3cab-4323-8637-7ff6dffce79b. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1106.150332] env[62385]: DEBUG oslo_concurrency.lockutils [req-f82e4972-30f2-4dd4-82ca-6a2444fdeb86 req-a6318424-0f46-46c6-bb13-eb9fb05ff816 service nova] Acquiring lock "refresh_cache-4fe375be-ba60-4601-bd89-610e12742df3" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1106.412024] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Releasing lock "refresh_cache-4fe375be-ba60-4601-bd89-610e12742df3" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1106.412377] env[62385]: DEBUG nova.compute.manager [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Instance network_info: |[{"id": "1e16fa5f-3cab-4323-8637-7ff6dffce79b", "address": "fa:16:3e:6c:0c:dd", "network": {"id": "9c4d6464-cacc-44f8-b7e5-0699cd312c53", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1889153615-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09c2c5e938bf46fd9353e47aecc72478", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4d3f69a-b086-4c3b-b976-5a848b63dfc4", "external-id": "nsx-vlan-transportzone-627", "segmentation_id": 627, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e16fa5f-3c", "ovs_interfaceid": "1e16fa5f-3cab-4323-8637-7ff6dffce79b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1106.412853] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Releasing lock "refresh_cache-1289e63a-5a9d-4615-b6cf-bbf5f4ae0317" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1106.416232] env[62385]: DEBUG oslo_concurrency.lockutils [req-f82e4972-30f2-4dd4-82ca-6a2444fdeb86 req-a6318424-0f46-46c6-bb13-eb9fb05ff816 service nova] Acquired lock "refresh_cache-4fe375be-ba60-4601-bd89-610e12742df3" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1106.416424] env[62385]: DEBUG nova.network.neutron [req-f82e4972-30f2-4dd4-82ca-6a2444fdeb86 req-a6318424-0f46-46c6-bb13-eb9fb05ff816 service nova] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Refreshing network info cache for port 1e16fa5f-3cab-4323-8637-7ff6dffce79b {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1106.417974] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6c:0c:dd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c4d3f69a-b086-4c3b-b976-5a848b63dfc4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1e16fa5f-3cab-4323-8637-7ff6dffce79b', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1106.425606] env[62385]: DEBUG oslo.service.loopingcall [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1106.428557] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1106.429017] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-69ceeac4-1923-4e25-b963-946f02af2739 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.452701] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1106.452701] env[62385]: value = "task-1206503" [ 1106.452701] env[62385]: _type = "Task" [ 1106.452701] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.461919] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206503, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.732318] env[62385]: DEBUG nova.network.neutron [req-f82e4972-30f2-4dd4-82ca-6a2444fdeb86 req-a6318424-0f46-46c6-bb13-eb9fb05ff816 service nova] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Updated VIF entry in instance network info cache for port 1e16fa5f-3cab-4323-8637-7ff6dffce79b. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1106.732778] env[62385]: DEBUG nova.network.neutron [req-f82e4972-30f2-4dd4-82ca-6a2444fdeb86 req-a6318424-0f46-46c6-bb13-eb9fb05ff816 service nova] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Updating instance_info_cache with network_info: [{"id": "1e16fa5f-3cab-4323-8637-7ff6dffce79b", "address": "fa:16:3e:6c:0c:dd", "network": {"id": "9c4d6464-cacc-44f8-b7e5-0699cd312c53", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1889153615-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09c2c5e938bf46fd9353e47aecc72478", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4d3f69a-b086-4c3b-b976-5a848b63dfc4", "external-id": "nsx-vlan-transportzone-627", "segmentation_id": 627, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e16fa5f-3c", "ovs_interfaceid": "1e16fa5f-3cab-4323-8637-7ff6dffce79b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1106.816859] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1106.817152] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1106.817363] env[62385]: DEBUG nova.objects.instance [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Lazy-loading 'pci_requests' on Instance uuid c49d4366-2ea0-44f2-8463-daf62bd1c40c {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1106.945026] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60524d84-9036-46f0-bea7-cc77d7ab41e1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.948176] env[62385]: DEBUG oslo_concurrency.lockutils [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "c4f0482d-5b46-4c28-9c40-a59b6aee7dc3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1106.948416] env[62385]: DEBUG oslo_concurrency.lockutils [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "c4f0482d-5b46-4c28-9c40-a59b6aee7dc3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1106.948620] env[62385]: DEBUG oslo_concurrency.lockutils [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "c4f0482d-5b46-4c28-9c40-a59b6aee7dc3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1106.948804] env[62385]: DEBUG oslo_concurrency.lockutils [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "c4f0482d-5b46-4c28-9c40-a59b6aee7dc3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1106.948972] env[62385]: DEBUG oslo_concurrency.lockutils [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "c4f0482d-5b46-4c28-9c40-a59b6aee7dc3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1106.963830] env[62385]: INFO nova.compute.manager [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Terminating instance [ 1106.968015] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-367b25d5-ced3-4fc5-b8bf-02876dee8ec1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.971237] env[62385]: DEBUG nova.compute.manager [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1106.971484] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1106.971707] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2d19d093-8626-4613-8a75-d35334707a6c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.977400] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206503, 'name': CreateVM_Task, 'duration_secs': 0.313261} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.980156] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1106.980462] env[62385]: DEBUG oslo_vmware.api [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1106.980462] env[62385]: value = "task-1206504" [ 1106.980462] env[62385]: _type = "Task" [ 1106.980462] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.980966] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Updating instance '1289e63a-5a9d-4615-b6cf-bbf5f4ae0317' progress to 83 {{(pid=62385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1106.984535] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1106.984696] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1106.985023] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1106.985448] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c9eb4d9-8f6e-4525-a826-d83f9375da7d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.993556] env[62385]: DEBUG oslo_vmware.api [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1106.993556] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521e5310-551e-16ac-2e20-85ae2d4f2dca" [ 1106.993556] env[62385]: _type = "Task" [ 1106.993556] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.999146] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] VM already powered off {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1106.999389] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Volume detach. Driver type: vmdk {{(pid=62385) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1106.999581] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261311', 'volume_id': '6c3e125e-0b69-44e6-9e1a-3527fedf6e2a', 'name': 'volume-6c3e125e-0b69-44e6-9e1a-3527fedf6e2a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c4f0482d-5b46-4c28-9c40-a59b6aee7dc3', 'attached_at': '', 'detached_at': '', 'volume_id': '6c3e125e-0b69-44e6-9e1a-3527fedf6e2a', 'serial': '6c3e125e-0b69-44e6-9e1a-3527fedf6e2a'} {{(pid=62385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1107.000280] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afda1777-b2a4-424c-a776-7f73aafc7a6a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.007483] env[62385]: DEBUG oslo_vmware.api [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521e5310-551e-16ac-2e20-85ae2d4f2dca, 'name': SearchDatastore_Task, 'duration_secs': 0.00868} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.020271] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1107.020505] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1107.020741] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1107.020894] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1107.021088] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1107.021704] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5bc67103-0662-4ca8-a5f3-534710bb95cb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.023789] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8004f04e-8c84-49c6-9b64-9aef9f533e46 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.031064] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ff91fca-8ed4-42ca-bc7f-e99cec3b2d94 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.034232] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1107.034409] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1107.035366] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91e86139-10b8-4cd9-9d40-b3c21213ae6a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.051212] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73e39e76-8702-4d3f-b47a-167b0ba433dc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.054532] env[62385]: DEBUG oslo_vmware.api [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1107.054532] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5200c820-ded5-71f8-1846-89a2b468dbc1" [ 1107.054532] env[62385]: _type = "Task" [ 1107.054532] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.067438] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] The volume has not been displaced from its original location: [datastore2] volume-6c3e125e-0b69-44e6-9e1a-3527fedf6e2a/volume-6c3e125e-0b69-44e6-9e1a-3527fedf6e2a.vmdk. No consolidation needed. {{(pid=62385) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1107.072647] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Reconfiguring VM instance instance-00000070 to detach disk 2000 {{(pid=62385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1107.073180] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1f2dbda7-4f45-49b9-aba8-1b45222454c0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.088904] env[62385]: DEBUG oslo_vmware.api [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5200c820-ded5-71f8-1846-89a2b468dbc1, 'name': SearchDatastore_Task, 'duration_secs': 0.007895} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.089917] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5fe5360c-7730-4405-9853-5a3296407a93 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.093946] env[62385]: DEBUG oslo_vmware.api [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1107.093946] env[62385]: value = "task-1206505" [ 1107.093946] env[62385]: _type = "Task" [ 1107.093946] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.095046] env[62385]: DEBUG oslo_vmware.api [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1107.095046] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5212b149-419a-f732-00eb-6abd48b1d87d" [ 1107.095046] env[62385]: _type = "Task" [ 1107.095046] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.104591] env[62385]: DEBUG oslo_vmware.api [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206505, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.107749] env[62385]: DEBUG oslo_vmware.api [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5212b149-419a-f732-00eb-6abd48b1d87d, 'name': SearchDatastore_Task, 'duration_secs': 0.007494} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.107974] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1107.108245] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 4fe375be-ba60-4601-bd89-610e12742df3/4fe375be-ba60-4601-bd89-610e12742df3.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1107.108469] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a30db8b7-39c3-450f-a328-517046813dc3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.114361] env[62385]: DEBUG oslo_vmware.api [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1107.114361] env[62385]: value = "task-1206506" [ 1107.114361] env[62385]: _type = "Task" [ 1107.114361] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.123234] env[62385]: DEBUG oslo_vmware.api [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206506, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.235669] env[62385]: DEBUG oslo_concurrency.lockutils [req-f82e4972-30f2-4dd4-82ca-6a2444fdeb86 req-a6318424-0f46-46c6-bb13-eb9fb05ff816 service nova] Releasing lock "refresh_cache-4fe375be-ba60-4601-bd89-610e12742df3" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1107.321768] env[62385]: DEBUG nova.objects.instance [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Lazy-loading 'numa_topology' on Instance uuid c49d4366-2ea0-44f2-8463-daf62bd1c40c {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1107.399165] env[62385]: DEBUG oslo_concurrency.lockutils [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1107.487510] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1107.487979] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1609f71d-55f6-4276-8976-092f8fc8889a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.498108] env[62385]: DEBUG oslo_vmware.api [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1107.498108] env[62385]: value = "task-1206507" [ 1107.498108] env[62385]: _type = "Task" [ 1107.498108] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.506396] env[62385]: DEBUG oslo_vmware.api [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206507, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.603730] env[62385]: DEBUG oslo_vmware.api [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206505, 'name': ReconfigVM_Task, 'duration_secs': 0.173753} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.604014] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Reconfigured VM instance instance-00000070 to detach disk 2000 {{(pid=62385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1107.608826] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b890f4af-8b50-40fe-953b-5bcb9c683d39 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.628237] env[62385]: DEBUG oslo_vmware.api [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206506, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.439474} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.629479] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 4fe375be-ba60-4601-bd89-610e12742df3/4fe375be-ba60-4601-bd89-610e12742df3.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1107.629723] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1107.630078] env[62385]: DEBUG oslo_vmware.api [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1107.630078] env[62385]: value = "task-1206508" [ 1107.630078] env[62385]: _type = "Task" [ 1107.630078] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.630299] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7fc95b5d-e7b2-4524-a83e-852dbb6ecf73 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.639289] env[62385]: DEBUG oslo_vmware.api [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206508, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.640402] env[62385]: DEBUG oslo_vmware.api [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1107.640402] env[62385]: value = "task-1206509" [ 1107.640402] env[62385]: _type = "Task" [ 1107.640402] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.646973] env[62385]: DEBUG oslo_vmware.api [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206509, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.823994] env[62385]: INFO nova.compute.claims [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1108.008400] env[62385]: DEBUG oslo_vmware.api [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206507, 'name': PowerOnVM_Task, 'duration_secs': 0.395794} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.008700] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1108.008907] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-1df2dbac-0ac8-4746-8044-f11b313e2196 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Updating instance '1289e63a-5a9d-4615-b6cf-bbf5f4ae0317' progress to 100 {{(pid=62385) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1108.141278] env[62385]: DEBUG oslo_vmware.api [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206508, 'name': ReconfigVM_Task, 'duration_secs': 0.117753} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.144205] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261311', 'volume_id': '6c3e125e-0b69-44e6-9e1a-3527fedf6e2a', 'name': 'volume-6c3e125e-0b69-44e6-9e1a-3527fedf6e2a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c4f0482d-5b46-4c28-9c40-a59b6aee7dc3', 'attached_at': '', 'detached_at': '', 'volume_id': '6c3e125e-0b69-44e6-9e1a-3527fedf6e2a', 'serial': '6c3e125e-0b69-44e6-9e1a-3527fedf6e2a'} {{(pid=62385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1108.144491] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1108.145213] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ee7191b-7590-437b-85ef-7a5585f6ecd2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.151559] env[62385]: DEBUG oslo_vmware.api [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206509, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083518} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.153404] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1108.153707] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1108.154371] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb1186a1-60c7-4865-a99e-6e2bc0c418bb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.156432] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f0d130b5-9bd3-4691-a8e5-c7eb9c0f8084 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.176234] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Reconfiguring VM instance instance-00000071 to attach disk [datastore1] 4fe375be-ba60-4601-bd89-610e12742df3/4fe375be-ba60-4601-bd89-610e12742df3.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1108.176447] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-58d19bdc-fa61-473e-b701-adbaae53e95b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.194528] env[62385]: DEBUG oslo_vmware.api [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1108.194528] env[62385]: value = "task-1206511" [ 1108.194528] env[62385]: _type = "Task" [ 1108.194528] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.203502] env[62385]: DEBUG oslo_vmware.api [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206511, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.220866] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1108.221119] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Deleting contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1108.221301] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Deleting the datastore file [datastore2] c4f0482d-5b46-4c28-9c40-a59b6aee7dc3 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1108.221664] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-24633743-cfe8-4989-9166-b9a61287df81 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.227074] env[62385]: DEBUG oslo_vmware.api [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1108.227074] env[62385]: value = "task-1206512" [ 1108.227074] env[62385]: _type = "Task" [ 1108.227074] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.234852] env[62385]: DEBUG oslo_vmware.api [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206512, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.285494] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0fe98998-8a22-4cfc-9688-71d42be29e5f tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "c4f0482d-5b46-4c28-9c40-a59b6aee7dc3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.704475] env[62385]: DEBUG oslo_vmware.api [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206511, 'name': ReconfigVM_Task, 'duration_secs': 0.259447} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.704837] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Reconfigured VM instance instance-00000071 to attach disk [datastore1] 4fe375be-ba60-4601-bd89-610e12742df3/4fe375be-ba60-4601-bd89-610e12742df3.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1108.705416] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8c1f1df2-4432-43e8-aa88-6eb64f32fb09 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.711668] env[62385]: DEBUG oslo_vmware.api [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1108.711668] env[62385]: value = "task-1206513" [ 1108.711668] env[62385]: _type = "Task" [ 1108.711668] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.719235] env[62385]: DEBUG oslo_vmware.api [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206513, 'name': Rename_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.735140] env[62385]: DEBUG oslo_vmware.api [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206512, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.174484} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.735409] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1108.735620] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Deleted contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1108.735803] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1108.735999] env[62385]: INFO nova.compute.manager [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Took 1.76 seconds to destroy the instance on the hypervisor. [ 1108.736260] env[62385]: DEBUG oslo.service.loopingcall [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1108.736464] env[62385]: DEBUG nova.compute.manager [-] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1108.736579] env[62385]: DEBUG nova.network.neutron [-] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1108.933723] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecac6be6-7045-4040-9445-132e64f55d6e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.941323] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7ea7e5e-5fc9-4f8e-a072-5f7ff94eae60 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.973036] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddbaf2f4-4718-4070-82d5-2a98eb645238 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.980823] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3cd7d18-9bfa-43ea-a576-95356c7a35e5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.995901] env[62385]: DEBUG nova.compute.provider_tree [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1109.222603] env[62385]: DEBUG oslo_vmware.api [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206513, 'name': Rename_Task, 'duration_secs': 0.167401} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.222603] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1109.222603] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d758959e-f82a-4089-90de-32e09bfa882e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.229029] env[62385]: DEBUG oslo_vmware.api [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1109.229029] env[62385]: value = "task-1206514" [ 1109.229029] env[62385]: _type = "Task" [ 1109.229029] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.236307] env[62385]: DEBUG oslo_vmware.api [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206514, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.298240] env[62385]: DEBUG nova.compute.manager [req-371e019f-11dc-48ef-8ef7-9dfbd8a810c7 req-e0788e67-5577-4d11-8d15-ba91f4203653 service nova] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Received event network-vif-deleted-07abbf5c-f2dc-4f07-a1a8-f81f59c6a2ae {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1109.298240] env[62385]: INFO nova.compute.manager [req-371e019f-11dc-48ef-8ef7-9dfbd8a810c7 req-e0788e67-5577-4d11-8d15-ba91f4203653 service nova] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Neutron deleted interface 07abbf5c-f2dc-4f07-a1a8-f81f59c6a2ae; detaching it from the instance and deleting it from the info cache [ 1109.298240] env[62385]: DEBUG nova.network.neutron [req-371e019f-11dc-48ef-8ef7-9dfbd8a810c7 req-e0788e67-5577-4d11-8d15-ba91f4203653 service nova] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1109.491754] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1109.499530] env[62385]: DEBUG nova.scheduler.client.report [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1109.636459] env[62385]: DEBUG oslo_concurrency.lockutils [None req-39ab1a40-bf65-48f5-a6ac-bf19a0b0aa14 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "5c9284ba-9f83-454f-b91b-9b0ea9ee3195" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1109.636751] env[62385]: DEBUG oslo_concurrency.lockutils [None req-39ab1a40-bf65-48f5-a6ac-bf19a0b0aa14 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "5c9284ba-9f83-454f-b91b-9b0ea9ee3195" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1109.637484] env[62385]: DEBUG oslo_concurrency.lockutils [None req-39ab1a40-bf65-48f5-a6ac-bf19a0b0aa14 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "5c9284ba-9f83-454f-b91b-9b0ea9ee3195-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1109.637704] env[62385]: DEBUG oslo_concurrency.lockutils [None req-39ab1a40-bf65-48f5-a6ac-bf19a0b0aa14 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "5c9284ba-9f83-454f-b91b-9b0ea9ee3195-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1109.637890] env[62385]: DEBUG oslo_concurrency.lockutils [None req-39ab1a40-bf65-48f5-a6ac-bf19a0b0aa14 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "5c9284ba-9f83-454f-b91b-9b0ea9ee3195-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.640221] env[62385]: INFO nova.compute.manager [None req-39ab1a40-bf65-48f5-a6ac-bf19a0b0aa14 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Terminating instance [ 1109.642118] env[62385]: DEBUG nova.compute.manager [None req-39ab1a40-bf65-48f5-a6ac-bf19a0b0aa14 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1109.642321] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-39ab1a40-bf65-48f5-a6ac-bf19a0b0aa14 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1109.643203] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e747e793-1f22-4eff-afab-2d551dddf133 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.650758] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-39ab1a40-bf65-48f5-a6ac-bf19a0b0aa14 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1109.650994] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7dfafd77-9fe7-4b36-af92-f7a814a09a2f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.657173] env[62385]: DEBUG oslo_vmware.api [None req-39ab1a40-bf65-48f5-a6ac-bf19a0b0aa14 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1109.657173] env[62385]: value = "task-1206515" [ 1109.657173] env[62385]: _type = "Task" [ 1109.657173] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.665370] env[62385]: DEBUG oslo_vmware.api [None req-39ab1a40-bf65-48f5-a6ac-bf19a0b0aa14 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206515, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.738391] env[62385]: DEBUG oslo_vmware.api [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206514, 'name': PowerOnVM_Task, 'duration_secs': 0.470832} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.738710] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1109.739341] env[62385]: INFO nova.compute.manager [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Took 5.31 seconds to spawn the instance on the hypervisor. [ 1109.739341] env[62385]: DEBUG nova.compute.manager [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1109.739900] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a763e20-dbac-4c7e-8301-fb59991885f1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.773057] env[62385]: DEBUG nova.network.neutron [-] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1109.802592] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c1812675-54ff-4416-ae49-cb238dbd1475 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.812879] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-304ba179-2f52-4067-a926-067d06e0dc80 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.840461] env[62385]: DEBUG nova.compute.manager [req-371e019f-11dc-48ef-8ef7-9dfbd8a810c7 req-e0788e67-5577-4d11-8d15-ba91f4203653 service nova] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Detach interface failed, port_id=07abbf5c-f2dc-4f07-a1a8-f81f59c6a2ae, reason: Instance c4f0482d-5b46-4c28-9c40-a59b6aee7dc3 could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1109.996665] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._sync_power_states {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1110.004095] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.187s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1110.006828] env[62385]: DEBUG oslo_concurrency.lockutils [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.608s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.056032] env[62385]: INFO nova.network.neutron [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Updating port b19d5597-14de-4b08-b628-3bc8e539f77a with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1110.138384] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81e82cb5-d75a-4f12-bf3e-f2932531ade3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.147899] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-701348c6-a22d-4b88-b7c7-a7d36ee92c4f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.186714] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da8b6b37-7b87-4743-80e5-55a9096057d0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.193957] env[62385]: DEBUG oslo_vmware.api [None req-39ab1a40-bf65-48f5-a6ac-bf19a0b0aa14 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206515, 'name': PowerOffVM_Task, 'duration_secs': 0.245346} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.195919] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-39ab1a40-bf65-48f5-a6ac-bf19a0b0aa14 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1110.196120] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-39ab1a40-bf65-48f5-a6ac-bf19a0b0aa14 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1110.196412] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-94b4a66e-904f-4e96-9199-22ff43147fcf {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.198669] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44cef76c-2649-4a1e-80b9-26e0fb61075f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.214457] env[62385]: DEBUG nova.compute.provider_tree [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1110.262237] env[62385]: INFO nova.compute.manager [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Took 10.04 seconds to build instance. [ 1110.268038] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-39ab1a40-bf65-48f5-a6ac-bf19a0b0aa14 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1110.268038] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-39ab1a40-bf65-48f5-a6ac-bf19a0b0aa14 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Deleting contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1110.268038] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-39ab1a40-bf65-48f5-a6ac-bf19a0b0aa14 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Deleting the datastore file [datastore2] 5c9284ba-9f83-454f-b91b-9b0ea9ee3195 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1110.268038] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-63be4e9f-34fd-450d-99ff-6a95f5b01678 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.272831] env[62385]: DEBUG oslo_vmware.api [None req-39ab1a40-bf65-48f5-a6ac-bf19a0b0aa14 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1110.272831] env[62385]: value = "task-1206517" [ 1110.272831] env[62385]: _type = "Task" [ 1110.272831] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.277445] env[62385]: INFO nova.compute.manager [-] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Took 1.54 seconds to deallocate network for instance. [ 1110.285301] env[62385]: DEBUG oslo_vmware.api [None req-39ab1a40-bf65-48f5-a6ac-bf19a0b0aa14 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206517, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.500353] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Getting list of instances from cluster (obj){ [ 1110.500353] env[62385]: value = "domain-c8" [ 1110.500353] env[62385]: _type = "ClusterComputeResource" [ 1110.500353] env[62385]: } {{(pid=62385) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 1110.501465] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67a26586-ef36-4093-92d6-b06e5e7282ed {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.514297] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Got total of 3 instances {{(pid=62385) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 1110.514478] env[62385]: WARNING nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] While synchronizing instance power states, found 6 instances in the database and 3 instances on the hypervisor. [ 1110.514596] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Triggering sync for uuid 9d6f098a-0b05-43ef-96b0-9eb99ad3538c {{(pid=62385) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1110.514784] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Triggering sync for uuid c49d4366-2ea0-44f2-8463-daf62bd1c40c {{(pid=62385) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1110.514935] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Triggering sync for uuid 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317 {{(pid=62385) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1110.515117] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Triggering sync for uuid 5c9284ba-9f83-454f-b91b-9b0ea9ee3195 {{(pid=62385) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1110.515266] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Triggering sync for uuid c4f0482d-5b46-4c28-9c40-a59b6aee7dc3 {{(pid=62385) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1110.515417] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Triggering sync for uuid 4fe375be-ba60-4601-bd89-610e12742df3 {{(pid=62385) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1110.515800] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquiring lock "9d6f098a-0b05-43ef-96b0-9eb99ad3538c" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.516047] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "9d6f098a-0b05-43ef-96b0-9eb99ad3538c" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.516317] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquiring lock "c49d4366-2ea0-44f2-8463-daf62bd1c40c" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.516534] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquiring lock "1289e63a-5a9d-4615-b6cf-bbf5f4ae0317" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.516717] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "1289e63a-5a9d-4615-b6cf-bbf5f4ae0317" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.516892] env[62385]: INFO nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] During sync_power_state the instance has a pending task (resize_reverting). Skip. [ 1110.517078] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "1289e63a-5a9d-4615-b6cf-bbf5f4ae0317" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1110.517279] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquiring lock "5c9284ba-9f83-454f-b91b-9b0ea9ee3195" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.517489] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquiring lock "c4f0482d-5b46-4c28-9c40-a59b6aee7dc3" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.517682] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquiring lock "4fe375be-ba60-4601-bd89-610e12742df3" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.518538] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c5651bc-015c-4447-9db4-4854b7202cb5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.573221] env[62385]: DEBUG nova.network.neutron [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Port c6ede9da-b209-4fae-b75a-c2f1ede4dcaf binding to destination host cpu-1 is already ACTIVE {{(pid=62385) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1110.573525] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "refresh_cache-1289e63a-5a9d-4615-b6cf-bbf5f4ae0317" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1110.573706] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquired lock "refresh_cache-1289e63a-5a9d-4615-b6cf-bbf5f4ae0317" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.574051] env[62385]: DEBUG nova.network.neutron [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1110.717217] env[62385]: DEBUG nova.scheduler.client.report [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1110.764984] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1aa8ec06-72ea-40ad-a7e0-59981532dd43 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lock "4fe375be-ba60-4601-bd89-610e12742df3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 11.548s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1110.765503] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "4fe375be-ba60-4601-bd89-610e12742df3" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.248s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.766180] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ce3085d-91a4-4252-801d-e9c4d7dd6f4f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.781851] env[62385]: DEBUG oslo_vmware.api [None req-39ab1a40-bf65-48f5-a6ac-bf19a0b0aa14 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206517, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.178752} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.782091] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-39ab1a40-bf65-48f5-a6ac-bf19a0b0aa14 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1110.782289] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-39ab1a40-bf65-48f5-a6ac-bf19a0b0aa14 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Deleted contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1110.782464] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-39ab1a40-bf65-48f5-a6ac-bf19a0b0aa14 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1110.782634] env[62385]: INFO nova.compute.manager [None req-39ab1a40-bf65-48f5-a6ac-bf19a0b0aa14 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1110.782865] env[62385]: DEBUG oslo.service.loopingcall [None req-39ab1a40-bf65-48f5-a6ac-bf19a0b0aa14 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1110.783059] env[62385]: DEBUG nova.compute.manager [-] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1110.783157] env[62385]: DEBUG nova.network.neutron [-] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1110.833749] env[62385]: INFO nova.compute.manager [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Took 0.56 seconds to detach 1 volumes for instance. [ 1110.835816] env[62385]: DEBUG nova.compute.manager [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Deleting volume: 6c3e125e-0b69-44e6-9e1a-3527fedf6e2a {{(pid=62385) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 1111.027398] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "9d6f098a-0b05-43ef-96b0-9eb99ad3538c" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.511s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.066365] env[62385]: DEBUG nova.compute.manager [req-125271a8-f580-4399-9e64-d2d9ca4b01c0 req-b1835987-ef66-4745-b4b9-bdd6b50e7ae3 service nova] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Received event network-vif-deleted-b5615be2-bc51-4fe3-8905-4b7083550949 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1111.066427] env[62385]: INFO nova.compute.manager [req-125271a8-f580-4399-9e64-d2d9ca4b01c0 req-b1835987-ef66-4745-b4b9-bdd6b50e7ae3 service nova] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Neutron deleted interface b5615be2-bc51-4fe3-8905-4b7083550949; detaching it from the instance and deleting it from the info cache [ 1111.066601] env[62385]: DEBUG nova.network.neutron [req-125271a8-f580-4399-9e64-d2d9ca4b01c0 req-b1835987-ef66-4745-b4b9-bdd6b50e7ae3 service nova] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1111.222491] env[62385]: DEBUG oslo_concurrency.lockutils [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.216s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.222726] env[62385]: INFO nova.compute.manager [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Successfully reverted task state from resize_migrating on failure for instance. [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server [None req-36669840-bf8c-4f1f-9fde-627336d2db0d tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Exception during message handling: AttributeError: 'NoneType' object has no attribute 'key' [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 172, in _process_incoming [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server raise self.value [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 166, in decorated_function [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server raise self.value [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 157, in decorated_function [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1453, in decorated_function [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 213, in decorated_function [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server raise self.value [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 203, in decorated_function [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 6071, in resize_instance [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server raise self.value [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 6068, in resize_instance [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server self._resize_instance(context, instance, image, migration, [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 6105, in _resize_instance [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server disk_info = self.driver.migrate_disk_and_power_off( [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 264, in migrate_disk_and_power_off [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server return self._vmops.migrate_disk_and_power_off(context, instance, [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1467, in migrate_disk_and_power_off [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server self._resize_disk(instance, vm_ref, vmdk, flavor) [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1398, in _resize_disk [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server self._volumeops.detach_disk_from_vm(vm_ref, instance, vmdk.device) [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 121, in detach_disk_from_vm [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server disk_key = device.key [ 1111.231243] env[62385]: ERROR oslo_messaging.rpc.server AttributeError: 'NoneType' object has no attribute 'key' [ 1111.232854] env[62385]: ERROR oslo_messaging.rpc.server [ 1111.278276] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "4fe375be-ba60-4601-bd89-610e12742df3" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.513s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.287090] env[62385]: DEBUG nova.network.neutron [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Updating instance_info_cache with network_info: [{"id": "c6ede9da-b209-4fae-b75a-c2f1ede4dcaf", "address": "fa:16:3e:a7:e1:c0", "network": {"id": "06017161-984c-4657-9639-4e6bdb6d4c09", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1072882119-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.208", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d45b5346eab44707b4d81eaaf12b0935", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39cd75b0-9ec7-48ed-b57f-34da0c573a60", "external-id": "nsx-vlan-transportzone-751", "segmentation_id": 751, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6ede9da-b2", "ovs_interfaceid": "c6ede9da-b209-4fae-b75a-c2f1ede4dcaf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1111.429166] env[62385]: DEBUG oslo_concurrency.lockutils [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.429479] env[62385]: DEBUG oslo_concurrency.lockutils [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.429736] env[62385]: DEBUG nova.objects.instance [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lazy-loading 'resources' on Instance uuid c4f0482d-5b46-4c28-9c40-a59b6aee7dc3 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1111.545638] env[62385]: DEBUG nova.network.neutron [-] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1111.569388] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-08283902-8eec-4e48-a324-5ae3ab266820 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.579480] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f139622-19d1-4cdd-9cac-fa57b100cd10 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.604045] env[62385]: DEBUG nova.compute.manager [req-125271a8-f580-4399-9e64-d2d9ca4b01c0 req-b1835987-ef66-4745-b4b9-bdd6b50e7ae3 service nova] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Detach interface failed, port_id=b5615be2-bc51-4fe3-8905-4b7083550949, reason: Instance 5c9284ba-9f83-454f-b91b-9b0ea9ee3195 could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1111.790124] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Releasing lock "refresh_cache-1289e63a-5a9d-4615-b6cf-bbf5f4ae0317" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1111.815933] env[62385]: DEBUG nova.compute.manager [req-28f9f8a0-61ee-40ec-973f-28b0b7c6447d req-9df39a4e-0d09-4037-be85-d894acaa67a7 service nova] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Received event network-vif-plugged-b19d5597-14de-4b08-b628-3bc8e539f77a {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1111.816176] env[62385]: DEBUG oslo_concurrency.lockutils [req-28f9f8a0-61ee-40ec-973f-28b0b7c6447d req-9df39a4e-0d09-4037-be85-d894acaa67a7 service nova] Acquiring lock "c49d4366-2ea0-44f2-8463-daf62bd1c40c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.816385] env[62385]: DEBUG oslo_concurrency.lockutils [req-28f9f8a0-61ee-40ec-973f-28b0b7c6447d req-9df39a4e-0d09-4037-be85-d894acaa67a7 service nova] Lock "c49d4366-2ea0-44f2-8463-daf62bd1c40c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.816557] env[62385]: DEBUG oslo_concurrency.lockutils [req-28f9f8a0-61ee-40ec-973f-28b0b7c6447d req-9df39a4e-0d09-4037-be85-d894acaa67a7 service nova] Lock "c49d4366-2ea0-44f2-8463-daf62bd1c40c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.816725] env[62385]: DEBUG nova.compute.manager [req-28f9f8a0-61ee-40ec-973f-28b0b7c6447d req-9df39a4e-0d09-4037-be85-d894acaa67a7 service nova] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] No waiting events found dispatching network-vif-plugged-b19d5597-14de-4b08-b628-3bc8e539f77a {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1111.816896] env[62385]: WARNING nova.compute.manager [req-28f9f8a0-61ee-40ec-973f-28b0b7c6447d req-9df39a4e-0d09-4037-be85-d894acaa67a7 service nova] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Received unexpected event network-vif-plugged-b19d5597-14de-4b08-b628-3bc8e539f77a for instance with vm_state shelved_offloaded and task_state spawning. [ 1111.996239] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Acquiring lock "refresh_cache-c49d4366-2ea0-44f2-8463-daf62bd1c40c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1111.996239] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Acquired lock "refresh_cache-c49d4366-2ea0-44f2-8463-daf62bd1c40c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1111.996346] env[62385]: DEBUG nova.network.neutron [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1112.045615] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ab28fda-1a62-49f4-8090-83d0829f107e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.048325] env[62385]: INFO nova.compute.manager [-] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Took 1.27 seconds to deallocate network for instance. [ 1112.056428] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f62ab5d-965c-4267-8006-561ae04fd991 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.088507] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d55f2c09-1ae9-44c1-a7b8-6c55c0154873 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.096542] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3dd309c-354a-447b-9899-f6b4c79de853 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.110041] env[62385]: DEBUG nova.compute.provider_tree [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1112.293873] env[62385]: DEBUG nova.compute.manager [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=62385) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1112.294145] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1112.556950] env[62385]: DEBUG oslo_concurrency.lockutils [None req-39ab1a40-bf65-48f5-a6ac-bf19a0b0aa14 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1112.612508] env[62385]: DEBUG nova.scheduler.client.report [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1112.801170] env[62385]: DEBUG nova.network.neutron [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Updating instance_info_cache with network_info: [{"id": "b19d5597-14de-4b08-b628-3bc8e539f77a", "address": "fa:16:3e:83:b5:c7", "network": {"id": "a561a523-c9d8-4169-b449-5c0bdef7b53b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-751953866-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d716d19f79945db98f684df0cf4b302", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb19d5597-14", "ovs_interfaceid": "b19d5597-14de-4b08-b628-3bc8e539f77a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1113.117696] env[62385]: DEBUG oslo_concurrency.lockutils [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.688s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.120584] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.826s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.140808] env[62385]: INFO nova.scheduler.client.report [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Deleted allocations for instance c4f0482d-5b46-4c28-9c40-a59b6aee7dc3 [ 1113.304390] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Releasing lock "refresh_cache-c49d4366-2ea0-44f2-8463-daf62bd1c40c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1113.330458] env[62385]: DEBUG nova.virt.hardware [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='7f65bada30179e5e5dec0d84d31712da',container_format='bare',created_at=2024-10-09T02:27:16Z,direct_url=,disk_format='vmdk',id=651910de-8449-4e93-bd88-46a95628369c,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-1985383717-shelved',owner='5d716d19f79945db98f684df0cf4b302',properties=ImageMetaProps,protected=,size=31659008,status='active',tags=,updated_at=2024-10-09T02:27:29Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1113.330705] env[62385]: DEBUG nova.virt.hardware [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1113.330867] env[62385]: DEBUG nova.virt.hardware [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1113.331070] env[62385]: DEBUG nova.virt.hardware [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1113.331228] env[62385]: DEBUG nova.virt.hardware [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1113.331382] env[62385]: DEBUG nova.virt.hardware [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1113.331684] env[62385]: DEBUG nova.virt.hardware [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1113.331853] env[62385]: DEBUG nova.virt.hardware [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1113.332038] env[62385]: DEBUG nova.virt.hardware [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1113.332344] env[62385]: DEBUG nova.virt.hardware [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1113.332600] env[62385]: DEBUG nova.virt.hardware [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1113.333474] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3800540-91e1-4f37-9576-930b0f877492 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.341255] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a992a86-9dde-4fb3-a591-d525b37b1375 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.353909] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:83:b5:c7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '170f3b82-5915-4e36-bce9-4664ebb6be5e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b19d5597-14de-4b08-b628-3bc8e539f77a', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1113.361122] env[62385]: DEBUG oslo.service.loopingcall [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1113.361346] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1113.361566] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-927669e1-9b51-4ab7-9264-17a65ad3e03c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.379462] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1113.379462] env[62385]: value = "task-1206519" [ 1113.379462] env[62385]: _type = "Task" [ 1113.379462] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.386281] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206519, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.623761] env[62385]: DEBUG nova.objects.instance [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lazy-loading 'migration_context' on Instance uuid 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1113.648718] env[62385]: DEBUG oslo_concurrency.lockutils [None req-55692d4c-33f3-48d3-868b-482ad77d0c35 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "c4f0482d-5b46-4c28-9c40-a59b6aee7dc3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.700s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.650059] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0fe98998-8a22-4cfc-9688-71d42be29e5f tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "c4f0482d-5b46-4c28-9c40-a59b6aee7dc3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 5.364s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.650059] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0fe98998-8a22-4cfc-9688-71d42be29e5f tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "c4f0482d-5b46-4c28-9c40-a59b6aee7dc3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.650306] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0fe98998-8a22-4cfc-9688-71d42be29e5f tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "c4f0482d-5b46-4c28-9c40-a59b6aee7dc3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.650306] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0fe98998-8a22-4cfc-9688-71d42be29e5f tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "c4f0482d-5b46-4c28-9c40-a59b6aee7dc3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.652540] env[62385]: INFO nova.compute.manager [None req-0fe98998-8a22-4cfc-9688-71d42be29e5f tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Terminating instance [ 1113.654696] env[62385]: DEBUG nova.compute.manager [None req-0fe98998-8a22-4cfc-9688-71d42be29e5f tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1113.654990] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4676f6ed-a1cd-4d7f-86ef-c707b3ffd172 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.663794] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b535d60-4faa-4be7-81f0-255f22907468 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.689690] env[62385]: WARNING nova.virt.vmwareapi.driver [None req-0fe98998-8a22-4cfc-9688-71d42be29e5f tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance c4f0482d-5b46-4c28-9c40-a59b6aee7dc3 could not be found. [ 1113.689880] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0fe98998-8a22-4cfc-9688-71d42be29e5f tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1113.690177] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1b2ce50e-6642-40fd-b5e0-66447a0f4e31 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.698430] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33656eda-408b-43eb-ade8-6237283e7f3f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.723416] env[62385]: WARNING nova.virt.vmwareapi.vmops [None req-0fe98998-8a22-4cfc-9688-71d42be29e5f tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c4f0482d-5b46-4c28-9c40-a59b6aee7dc3 could not be found. [ 1113.723613] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-0fe98998-8a22-4cfc-9688-71d42be29e5f tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1113.723825] env[62385]: INFO nova.compute.manager [None req-0fe98998-8a22-4cfc-9688-71d42be29e5f tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Took 0.07 seconds to destroy the instance on the hypervisor. [ 1113.724092] env[62385]: DEBUG oslo.service.loopingcall [None req-0fe98998-8a22-4cfc-9688-71d42be29e5f tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1113.724324] env[62385]: DEBUG nova.compute.manager [-] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1113.724419] env[62385]: DEBUG nova.network.neutron [-] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1113.846145] env[62385]: DEBUG nova.compute.manager [req-5b242260-f1b3-43cd-9351-e6ff9e9ad53a req-b9120c01-0e43-4231-b87d-11b4a4b5e20f service nova] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Received event network-changed-1e16fa5f-3cab-4323-8637-7ff6dffce79b {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1113.846411] env[62385]: DEBUG nova.compute.manager [req-5b242260-f1b3-43cd-9351-e6ff9e9ad53a req-b9120c01-0e43-4231-b87d-11b4a4b5e20f service nova] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Refreshing instance network info cache due to event network-changed-1e16fa5f-3cab-4323-8637-7ff6dffce79b. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1113.846657] env[62385]: DEBUG oslo_concurrency.lockutils [req-5b242260-f1b3-43cd-9351-e6ff9e9ad53a req-b9120c01-0e43-4231-b87d-11b4a4b5e20f service nova] Acquiring lock "refresh_cache-4fe375be-ba60-4601-bd89-610e12742df3" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1113.846752] env[62385]: DEBUG oslo_concurrency.lockutils [req-5b242260-f1b3-43cd-9351-e6ff9e9ad53a req-b9120c01-0e43-4231-b87d-11b4a4b5e20f service nova] Acquired lock "refresh_cache-4fe375be-ba60-4601-bd89-610e12742df3" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1113.846864] env[62385]: DEBUG nova.network.neutron [req-5b242260-f1b3-43cd-9351-e6ff9e9ad53a req-b9120c01-0e43-4231-b87d-11b4a4b5e20f service nova] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Refreshing network info cache for port 1e16fa5f-3cab-4323-8637-7ff6dffce79b {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1113.889721] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206519, 'name': CreateVM_Task, 'duration_secs': 0.352126} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.889896] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1113.890594] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/651910de-8449-4e93-bd88-46a95628369c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1113.890787] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Acquired lock "[datastore1] devstack-image-cache_base/651910de-8449-4e93-bd88-46a95628369c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1113.891168] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/651910de-8449-4e93-bd88-46a95628369c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1113.891444] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e5b1d99b-5a44-40c7-a757-4e4d496bf64f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.897257] env[62385]: DEBUG oslo_vmware.api [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for the task: (returnval){ [ 1113.897257] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]525c547e-4e5d-4dd9-cbcd-7f7637f8f630" [ 1113.897257] env[62385]: _type = "Task" [ 1113.897257] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.905317] env[62385]: DEBUG oslo_vmware.api [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]525c547e-4e5d-4dd9-cbcd-7f7637f8f630, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.205524] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4502522a-2207-4610-8243-39361512f876 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.213176] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8dcedb9-7b18-481b-8008-26866facda02 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.242583] env[62385]: DEBUG nova.network.neutron [-] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1114.244361] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a753390f-7ad7-49e0-b88b-9210028665e5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.251498] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00ecced5-69f5-4ace-9fcc-f0b352625d8a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.264743] env[62385]: DEBUG nova.compute.provider_tree [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1114.408664] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Releasing lock "[datastore1] devstack-image-cache_base/651910de-8449-4e93-bd88-46a95628369c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1114.408882] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Processing image 651910de-8449-4e93-bd88-46a95628369c {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1114.409135] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/651910de-8449-4e93-bd88-46a95628369c/651910de-8449-4e93-bd88-46a95628369c.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1114.409292] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Acquired lock "[datastore1] devstack-image-cache_base/651910de-8449-4e93-bd88-46a95628369c/651910de-8449-4e93-bd88-46a95628369c.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1114.409481] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1114.409736] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0b64ce68-6b35-443c-a751-7476d0c03d42 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.418363] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1114.418611] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1114.419351] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b337d78a-78ec-4560-bda7-61b5f946f318 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.424552] env[62385]: DEBUG oslo_vmware.api [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for the task: (returnval){ [ 1114.424552] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521e512f-1b1d-c6f0-10d5-3e5998596fa6" [ 1114.424552] env[62385]: _type = "Task" [ 1114.424552] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.433219] env[62385]: DEBUG oslo_vmware.api [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]521e512f-1b1d-c6f0-10d5-3e5998596fa6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.558077] env[62385]: DEBUG nova.network.neutron [req-5b242260-f1b3-43cd-9351-e6ff9e9ad53a req-b9120c01-0e43-4231-b87d-11b4a4b5e20f service nova] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Updated VIF entry in instance network info cache for port 1e16fa5f-3cab-4323-8637-7ff6dffce79b. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1114.558591] env[62385]: DEBUG nova.network.neutron [req-5b242260-f1b3-43cd-9351-e6ff9e9ad53a req-b9120c01-0e43-4231-b87d-11b4a4b5e20f service nova] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Updating instance_info_cache with network_info: [{"id": "1e16fa5f-3cab-4323-8637-7ff6dffce79b", "address": "fa:16:3e:6c:0c:dd", "network": {"id": "9c4d6464-cacc-44f8-b7e5-0699cd312c53", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1889153615-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.148", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09c2c5e938bf46fd9353e47aecc72478", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4d3f69a-b086-4c3b-b976-5a848b63dfc4", "external-id": "nsx-vlan-transportzone-627", "segmentation_id": 627, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e16fa5f-3c", "ovs_interfaceid": "1e16fa5f-3cab-4323-8637-7ff6dffce79b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1114.747885] env[62385]: INFO nova.compute.manager [-] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Took 1.02 seconds to deallocate network for instance. [ 1114.762701] env[62385]: WARNING nova.volume.cinder [None req-0fe98998-8a22-4cfc-9688-71d42be29e5f tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Attachment 7d66210e-294d-4be3-952b-17c2621eb2a6 does not exist. Ignoring.: cinderclient.exceptions.NotFound: Volume attachment could not be found with filter: attachment_id = 7d66210e-294d-4be3-952b-17c2621eb2a6. (HTTP 404) (Request-ID: req-5edc413f-ef5c-42c8-9ba3-3e93b4d4f981) [ 1114.763010] env[62385]: INFO nova.compute.manager [None req-0fe98998-8a22-4cfc-9688-71d42be29e5f tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Took 0.01 seconds to detach 1 volumes for instance. [ 1114.765323] env[62385]: DEBUG nova.compute.manager [None req-0fe98998-8a22-4cfc-9688-71d42be29e5f tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Deleting volume: 6c3e125e-0b69-44e6-9e1a-3527fedf6e2a {{(pid=62385) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 1114.767784] env[62385]: DEBUG nova.scheduler.client.report [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1114.783363] env[62385]: WARNING nova.compute.manager [None req-0fe98998-8a22-4cfc-9688-71d42be29e5f tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Failed to delete volume: 6c3e125e-0b69-44e6-9e1a-3527fedf6e2a due to Volume 6c3e125e-0b69-44e6-9e1a-3527fedf6e2a could not be found.: nova.exception.VolumeNotFound: Volume 6c3e125e-0b69-44e6-9e1a-3527fedf6e2a could not be found. [ 1114.937029] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Preparing fetch location {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1114.937029] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Fetch image to [datastore1] OSTACK_IMG_1c3a3b5b-1e1b-4150-b704-be3965cddd32/OSTACK_IMG_1c3a3b5b-1e1b-4150-b704-be3965cddd32.vmdk {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1114.937363] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Downloading stream optimized image 651910de-8449-4e93-bd88-46a95628369c to [datastore1] OSTACK_IMG_1c3a3b5b-1e1b-4150-b704-be3965cddd32/OSTACK_IMG_1c3a3b5b-1e1b-4150-b704-be3965cddd32.vmdk on the data store datastore1 as vApp {{(pid=62385) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1114.937363] env[62385]: DEBUG nova.virt.vmwareapi.images [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Downloading image file data 651910de-8449-4e93-bd88-46a95628369c to the ESX as VM named 'OSTACK_IMG_1c3a3b5b-1e1b-4150-b704-be3965cddd32' {{(pid=62385) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1115.006921] env[62385]: DEBUG oslo_vmware.rw_handles [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1115.006921] env[62385]: value = "resgroup-9" [ 1115.006921] env[62385]: _type = "ResourcePool" [ 1115.006921] env[62385]: }. {{(pid=62385) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1115.007327] env[62385]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-4a42aeee-c804-4f2d-84c8-5709bb1c847a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.030671] env[62385]: DEBUG oslo_vmware.rw_handles [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Lease: (returnval){ [ 1115.030671] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]523dddf2-d2f1-fe2d-593b-eef90af395e9" [ 1115.030671] env[62385]: _type = "HttpNfcLease" [ 1115.030671] env[62385]: } obtained for vApp import into resource pool (val){ [ 1115.030671] env[62385]: value = "resgroup-9" [ 1115.030671] env[62385]: _type = "ResourcePool" [ 1115.030671] env[62385]: }. {{(pid=62385) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1115.030927] env[62385]: DEBUG oslo_vmware.api [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for the lease: (returnval){ [ 1115.030927] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]523dddf2-d2f1-fe2d-593b-eef90af395e9" [ 1115.030927] env[62385]: _type = "HttpNfcLease" [ 1115.030927] env[62385]: } to be ready. {{(pid=62385) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1115.037914] env[62385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1115.037914] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]523dddf2-d2f1-fe2d-593b-eef90af395e9" [ 1115.037914] env[62385]: _type = "HttpNfcLease" [ 1115.037914] env[62385]: } is initializing. {{(pid=62385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1115.061837] env[62385]: DEBUG oslo_concurrency.lockutils [req-5b242260-f1b3-43cd-9351-e6ff9e9ad53a req-b9120c01-0e43-4231-b87d-11b4a4b5e20f service nova] Releasing lock "refresh_cache-4fe375be-ba60-4601-bd89-610e12742df3" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1115.062134] env[62385]: DEBUG nova.compute.manager [req-5b242260-f1b3-43cd-9351-e6ff9e9ad53a req-b9120c01-0e43-4231-b87d-11b4a4b5e20f service nova] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Received event network-changed-b19d5597-14de-4b08-b628-3bc8e539f77a {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1115.062314] env[62385]: DEBUG nova.compute.manager [req-5b242260-f1b3-43cd-9351-e6ff9e9ad53a req-b9120c01-0e43-4231-b87d-11b4a4b5e20f service nova] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Refreshing instance network info cache due to event network-changed-b19d5597-14de-4b08-b628-3bc8e539f77a. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1115.062540] env[62385]: DEBUG oslo_concurrency.lockutils [req-5b242260-f1b3-43cd-9351-e6ff9e9ad53a req-b9120c01-0e43-4231-b87d-11b4a4b5e20f service nova] Acquiring lock "refresh_cache-c49d4366-2ea0-44f2-8463-daf62bd1c40c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1115.062685] env[62385]: DEBUG oslo_concurrency.lockutils [req-5b242260-f1b3-43cd-9351-e6ff9e9ad53a req-b9120c01-0e43-4231-b87d-11b4a4b5e20f service nova] Acquired lock "refresh_cache-c49d4366-2ea0-44f2-8463-daf62bd1c40c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1115.062890] env[62385]: DEBUG nova.network.neutron [req-5b242260-f1b3-43cd-9351-e6ff9e9ad53a req-b9120c01-0e43-4231-b87d-11b4a4b5e20f service nova] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Refreshing network info cache for port b19d5597-14de-4b08-b628-3bc8e539f77a {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1115.288320] env[62385]: INFO nova.compute.manager [None req-0fe98998-8a22-4cfc-9688-71d42be29e5f tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Instance disappeared during terminate [ 1115.288511] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0fe98998-8a22-4cfc-9688-71d42be29e5f tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "c4f0482d-5b46-4c28-9c40-a59b6aee7dc3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 1.639s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.289534] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "c4f0482d-5b46-4c28-9c40-a59b6aee7dc3" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 4.772s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.289731] env[62385]: INFO nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] During sync_power_state the instance has a pending task (deleting). Skip. [ 1115.289911] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "c4f0482d-5b46-4c28-9c40-a59b6aee7dc3" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.539273] env[62385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1115.539273] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]523dddf2-d2f1-fe2d-593b-eef90af395e9" [ 1115.539273] env[62385]: _type = "HttpNfcLease" [ 1115.539273] env[62385]: } is ready. {{(pid=62385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1115.539556] env[62385]: DEBUG oslo_vmware.rw_handles [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1115.539556] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]523dddf2-d2f1-fe2d-593b-eef90af395e9" [ 1115.539556] env[62385]: _type = "HttpNfcLease" [ 1115.539556] env[62385]: }. {{(pid=62385) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1115.540286] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ce69982-b600-46ad-9779-d6f4507f3a31 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.548061] env[62385]: DEBUG oslo_vmware.rw_handles [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e97c48-a4df-3db3-4f44-70159f2812bf/disk-0.vmdk from lease info. {{(pid=62385) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1115.548061] env[62385]: DEBUG oslo_vmware.rw_handles [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Creating HTTP connection to write to file with size = 31659008 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e97c48-a4df-3db3-4f44-70159f2812bf/disk-0.vmdk. {{(pid=62385) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1115.612206] env[62385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-f7109165-80c4-490e-85e0-55886235dbdb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.777726] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.657s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.783691] env[62385]: DEBUG oslo_concurrency.lockutils [None req-39ab1a40-bf65-48f5-a6ac-bf19a0b0aa14 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.227s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.783952] env[62385]: DEBUG nova.objects.instance [None req-39ab1a40-bf65-48f5-a6ac-bf19a0b0aa14 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lazy-loading 'resources' on Instance uuid 5c9284ba-9f83-454f-b91b-9b0ea9ee3195 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1115.802670] env[62385]: DEBUG nova.network.neutron [req-5b242260-f1b3-43cd-9351-e6ff9e9ad53a req-b9120c01-0e43-4231-b87d-11b4a4b5e20f service nova] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Updated VIF entry in instance network info cache for port b19d5597-14de-4b08-b628-3bc8e539f77a. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1115.803026] env[62385]: DEBUG nova.network.neutron [req-5b242260-f1b3-43cd-9351-e6ff9e9ad53a req-b9120c01-0e43-4231-b87d-11b4a4b5e20f service nova] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Updating instance_info_cache with network_info: [{"id": "b19d5597-14de-4b08-b628-3bc8e539f77a", "address": "fa:16:3e:83:b5:c7", "network": {"id": "a561a523-c9d8-4169-b449-5c0bdef7b53b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-751953866-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d716d19f79945db98f684df0cf4b302", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb19d5597-14", "ovs_interfaceid": "b19d5597-14de-4b08-b628-3bc8e539f77a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1116.305387] env[62385]: DEBUG oslo_concurrency.lockutils [req-5b242260-f1b3-43cd-9351-e6ff9e9ad53a req-b9120c01-0e43-4231-b87d-11b4a4b5e20f service nova] Releasing lock "refresh_cache-c49d4366-2ea0-44f2-8463-daf62bd1c40c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1116.377658] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81f597fe-dc63-47da-8bf5-bc83358d6338 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.389260] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5f0c3ce-1691-4d4b-8a43-407378e56225 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.428905] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d17b2fe-48cc-495b-8737-c85207511564 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.440709] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d50ded89-0439-4bc2-9350-e5d69198cfc0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.455708] env[62385]: DEBUG nova.compute.provider_tree [None req-39ab1a40-bf65-48f5-a6ac-bf19a0b0aa14 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1116.852666] env[62385]: DEBUG oslo_vmware.rw_handles [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Completed reading data from the image iterator. {{(pid=62385) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1116.852849] env[62385]: DEBUG oslo_vmware.rw_handles [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e97c48-a4df-3db3-4f44-70159f2812bf/disk-0.vmdk. {{(pid=62385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1116.853871] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d12464e9-c4a5-4f11-8d8e-82cb92196112 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.861154] env[62385]: DEBUG oslo_vmware.rw_handles [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e97c48-a4df-3db3-4f44-70159f2812bf/disk-0.vmdk is in state: ready. {{(pid=62385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1116.861296] env[62385]: DEBUG oslo_vmware.rw_handles [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e97c48-a4df-3db3-4f44-70159f2812bf/disk-0.vmdk. {{(pid=62385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1116.861486] env[62385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-643d0733-c842-4ad0-994c-bc3556babfc1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.960357] env[62385]: DEBUG nova.scheduler.client.report [None req-39ab1a40-bf65-48f5-a6ac-bf19a0b0aa14 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1117.032604] env[62385]: DEBUG oslo_vmware.rw_handles [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e97c48-a4df-3db3-4f44-70159f2812bf/disk-0.vmdk. {{(pid=62385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1117.033181] env[62385]: INFO nova.virt.vmwareapi.images [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Downloaded image file data 651910de-8449-4e93-bd88-46a95628369c [ 1117.033674] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaf2d255-dee8-404c-a5eb-90df14ea04a6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.048901] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7275ea5a-0089-4c19-a50c-c7257d89f64b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.074337] env[62385]: INFO nova.virt.vmwareapi.images [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] The imported VM was unregistered [ 1117.076623] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Caching image {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1117.076849] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Creating directory with path [datastore1] devstack-image-cache_base/651910de-8449-4e93-bd88-46a95628369c {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1117.077116] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e699cf33-0d05-44c7-853f-b8650af08a9a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.086871] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Created directory with path [datastore1] devstack-image-cache_base/651910de-8449-4e93-bd88-46a95628369c {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1117.087063] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_1c3a3b5b-1e1b-4150-b704-be3965cddd32/OSTACK_IMG_1c3a3b5b-1e1b-4150-b704-be3965cddd32.vmdk to [datastore1] devstack-image-cache_base/651910de-8449-4e93-bd88-46a95628369c/651910de-8449-4e93-bd88-46a95628369c.vmdk. {{(pid=62385) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1117.087289] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-846a046e-9e92-44f0-983d-35bf320fec74 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.093463] env[62385]: DEBUG oslo_vmware.api [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for the task: (returnval){ [ 1117.093463] env[62385]: value = "task-1206522" [ 1117.093463] env[62385]: _type = "Task" [ 1117.093463] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.100163] env[62385]: DEBUG oslo_vmware.api [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206522, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.316801] env[62385]: INFO nova.compute.manager [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Swapping old allocation on dict_keys(['8c6086eb-4f96-48ed-af58-d7d7f36e7c5a']) held by migration c894b6b6-f1f1-4d64-b8a8-2e573459e7f3 for instance [ 1117.343443] env[62385]: DEBUG nova.scheduler.client.report [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Overwriting current allocation {'allocations': {'8c6086eb-4f96-48ed-af58-d7d7f36e7c5a': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 148}}, 'project_id': 'd45b5346eab44707b4d81eaaf12b0935', 'user_id': '7e04c6591fe043d686140eb2f5ce0be2', 'consumer_generation': 1} on consumer 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317 {{(pid=62385) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1117.432139] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "refresh_cache-1289e63a-5a9d-4615-b6cf-bbf5f4ae0317" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1117.432458] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquired lock "refresh_cache-1289e63a-5a9d-4615-b6cf-bbf5f4ae0317" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1117.432670] env[62385]: DEBUG nova.network.neutron [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1117.465732] env[62385]: DEBUG oslo_concurrency.lockutils [None req-39ab1a40-bf65-48f5-a6ac-bf19a0b0aa14 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.682s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.486044] env[62385]: INFO nova.scheduler.client.report [None req-39ab1a40-bf65-48f5-a6ac-bf19a0b0aa14 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Deleted allocations for instance 5c9284ba-9f83-454f-b91b-9b0ea9ee3195 [ 1117.603446] env[62385]: DEBUG oslo_vmware.api [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206522, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.994471] env[62385]: DEBUG oslo_concurrency.lockutils [None req-39ab1a40-bf65-48f5-a6ac-bf19a0b0aa14 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "5c9284ba-9f83-454f-b91b-9b0ea9ee3195" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.358s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.995475] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "5c9284ba-9f83-454f-b91b-9b0ea9ee3195" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 7.478s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.995677] env[62385]: INFO nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] During sync_power_state the instance has a pending task (deleting). Skip. [ 1117.995864] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "5c9284ba-9f83-454f-b91b-9b0ea9ee3195" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1118.103322] env[62385]: DEBUG oslo_vmware.api [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206522, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.179591] env[62385]: DEBUG nova.network.neutron [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Updating instance_info_cache with network_info: [{"id": "c6ede9da-b209-4fae-b75a-c2f1ede4dcaf", "address": "fa:16:3e:a7:e1:c0", "network": {"id": "06017161-984c-4657-9639-4e6bdb6d4c09", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1072882119-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.208", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d45b5346eab44707b4d81eaaf12b0935", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39cd75b0-9ec7-48ed-b57f-34da0c573a60", "external-id": "nsx-vlan-transportzone-751", "segmentation_id": 751, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6ede9da-b2", "ovs_interfaceid": "c6ede9da-b209-4fae-b75a-c2f1ede4dcaf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1118.604181] env[62385]: DEBUG oslo_vmware.api [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206522, 'name': MoveVirtualDisk_Task} progress is 69%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.683152] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Releasing lock "refresh_cache-1289e63a-5a9d-4615-b6cf-bbf5f4ae0317" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1118.683667] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1118.683979] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-090cae6d-c14e-4df0-9614-48a2257117f3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.691787] env[62385]: DEBUG oslo_vmware.api [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1118.691787] env[62385]: value = "task-1206524" [ 1118.691787] env[62385]: _type = "Task" [ 1118.691787] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.700187] env[62385]: DEBUG oslo_vmware.api [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206524, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.106503] env[62385]: DEBUG oslo_vmware.api [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206522, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.201824] env[62385]: DEBUG oslo_vmware.api [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206524, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.606933] env[62385]: DEBUG oslo_vmware.api [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206522, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.24195} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.607328] env[62385]: INFO nova.virt.vmwareapi.ds_util [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_1c3a3b5b-1e1b-4150-b704-be3965cddd32/OSTACK_IMG_1c3a3b5b-1e1b-4150-b704-be3965cddd32.vmdk to [datastore1] devstack-image-cache_base/651910de-8449-4e93-bd88-46a95628369c/651910de-8449-4e93-bd88-46a95628369c.vmdk. [ 1119.607412] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Cleaning up location [datastore1] OSTACK_IMG_1c3a3b5b-1e1b-4150-b704-be3965cddd32 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1119.607565] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_1c3a3b5b-1e1b-4150-b704-be3965cddd32 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1119.607816] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b0d4f603-dcdb-47ee-ac4b-cf2ed8c4d714 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.613820] env[62385]: DEBUG oslo_vmware.api [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for the task: (returnval){ [ 1119.613820] env[62385]: value = "task-1206525" [ 1119.613820] env[62385]: _type = "Task" [ 1119.613820] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.621166] env[62385]: DEBUG oslo_vmware.api [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206525, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.701830] env[62385]: DEBUG oslo_vmware.api [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206524, 'name': PowerOffVM_Task, 'duration_secs': 0.620627} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.701968] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1119.702680] env[62385]: DEBUG nova.virt.hardware [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1119.702931] env[62385]: DEBUG nova.virt.hardware [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1119.703109] env[62385]: DEBUG nova.virt.hardware [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1119.703303] env[62385]: DEBUG nova.virt.hardware [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1119.703455] env[62385]: DEBUG nova.virt.hardware [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1119.703609] env[62385]: DEBUG nova.virt.hardware [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1119.703850] env[62385]: DEBUG nova.virt.hardware [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1119.704044] env[62385]: DEBUG nova.virt.hardware [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1119.704229] env[62385]: DEBUG nova.virt.hardware [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1119.704409] env[62385]: DEBUG nova.virt.hardware [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1119.704581] env[62385]: DEBUG nova.virt.hardware [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1119.709508] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2976b9fd-f67f-4321-94e8-d4f13da3d2dd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.724448] env[62385]: DEBUG oslo_vmware.api [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1119.724448] env[62385]: value = "task-1206526" [ 1119.724448] env[62385]: _type = "Task" [ 1119.724448] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.732102] env[62385]: DEBUG oslo_vmware.api [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206526, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.123434] env[62385]: DEBUG oslo_vmware.api [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206525, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.031286} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.123708] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1120.123874] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Releasing lock "[datastore1] devstack-image-cache_base/651910de-8449-4e93-bd88-46a95628369c/651910de-8449-4e93-bd88-46a95628369c.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1120.124148] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/651910de-8449-4e93-bd88-46a95628369c/651910de-8449-4e93-bd88-46a95628369c.vmdk to [datastore1] c49d4366-2ea0-44f2-8463-daf62bd1c40c/c49d4366-2ea0-44f2-8463-daf62bd1c40c.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1120.124402] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a2adffe7-54ba-4c6c-b68c-4bf31aa6f6c5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.130331] env[62385]: DEBUG oslo_vmware.api [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for the task: (returnval){ [ 1120.130331] env[62385]: value = "task-1206527" [ 1120.130331] env[62385]: _type = "Task" [ 1120.130331] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.138393] env[62385]: DEBUG oslo_vmware.api [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206527, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.234148] env[62385]: DEBUG oslo_vmware.api [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206526, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.640818] env[62385]: DEBUG oslo_vmware.api [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206527, 'name': CopyVirtualDisk_Task} progress is 15%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.734925] env[62385]: DEBUG oslo_vmware.api [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206526, 'name': ReconfigVM_Task, 'duration_secs': 0.679749} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.735797] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-018ff004-a684-49fd-b50d-eee8b6686952 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.754414] env[62385]: DEBUG nova.virt.hardware [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1120.754712] env[62385]: DEBUG nova.virt.hardware [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1120.754832] env[62385]: DEBUG nova.virt.hardware [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1120.755060] env[62385]: DEBUG nova.virt.hardware [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1120.755236] env[62385]: DEBUG nova.virt.hardware [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1120.755391] env[62385]: DEBUG nova.virt.hardware [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1120.755608] env[62385]: DEBUG nova.virt.hardware [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1120.755974] env[62385]: DEBUG nova.virt.hardware [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1120.755974] env[62385]: DEBUG nova.virt.hardware [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1120.756105] env[62385]: DEBUG nova.virt.hardware [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1120.756284] env[62385]: DEBUG nova.virt.hardware [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1120.757146] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ecc42d7e-c207-4b80-9a8b-f4b77a665868 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.762712] env[62385]: DEBUG oslo_vmware.api [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1120.762712] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5227e0b8-47a1-451e-21b9-31e9fe81bc11" [ 1120.762712] env[62385]: _type = "Task" [ 1120.762712] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.772045] env[62385]: DEBUG oslo_vmware.api [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5227e0b8-47a1-451e-21b9-31e9fe81bc11, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.141769] env[62385]: DEBUG oslo_vmware.api [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206527, 'name': CopyVirtualDisk_Task} progress is 38%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.280093] env[62385]: DEBUG oslo_vmware.api [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5227e0b8-47a1-451e-21b9-31e9fe81bc11, 'name': SearchDatastore_Task, 'duration_secs': 0.046248} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.290796] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Reconfiguring VM instance instance-0000006c to detach disk 2000 {{(pid=62385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1121.292512] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-22ddb216-132a-4c35-b07d-7a07d55c1132 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.326071] env[62385]: DEBUG oslo_vmware.api [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1121.326071] env[62385]: value = "task-1206528" [ 1121.326071] env[62385]: _type = "Task" [ 1121.326071] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.340748] env[62385]: DEBUG oslo_vmware.api [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206528, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.551158] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1121.643164] env[62385]: DEBUG oslo_vmware.api [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206527, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.757139] env[62385]: DEBUG oslo_concurrency.lockutils [None req-988cd0f7-e95a-4e74-8866-a40a65d76147 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "9d6f098a-0b05-43ef-96b0-9eb99ad3538c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1121.757435] env[62385]: DEBUG oslo_concurrency.lockutils [None req-988cd0f7-e95a-4e74-8866-a40a65d76147 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "9d6f098a-0b05-43ef-96b0-9eb99ad3538c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1121.757663] env[62385]: DEBUG oslo_concurrency.lockutils [None req-988cd0f7-e95a-4e74-8866-a40a65d76147 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "9d6f098a-0b05-43ef-96b0-9eb99ad3538c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1121.757855] env[62385]: DEBUG oslo_concurrency.lockutils [None req-988cd0f7-e95a-4e74-8866-a40a65d76147 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "9d6f098a-0b05-43ef-96b0-9eb99ad3538c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1121.758049] env[62385]: DEBUG oslo_concurrency.lockutils [None req-988cd0f7-e95a-4e74-8866-a40a65d76147 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "9d6f098a-0b05-43ef-96b0-9eb99ad3538c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1121.760399] env[62385]: INFO nova.compute.manager [None req-988cd0f7-e95a-4e74-8866-a40a65d76147 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Terminating instance [ 1121.762478] env[62385]: DEBUG nova.compute.manager [None req-988cd0f7-e95a-4e74-8866-a40a65d76147 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1121.762707] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-988cd0f7-e95a-4e74-8866-a40a65d76147 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1121.763558] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4a26fe5-712e-4423-afec-4845bb36309f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.771829] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-988cd0f7-e95a-4e74-8866-a40a65d76147 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1121.772156] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3a893442-1bff-428b-9275-7c00953324e2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.778352] env[62385]: DEBUG oslo_vmware.api [None req-988cd0f7-e95a-4e74-8866-a40a65d76147 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1121.778352] env[62385]: value = "task-1206529" [ 1121.778352] env[62385]: _type = "Task" [ 1121.778352] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.786958] env[62385]: DEBUG oslo_vmware.api [None req-988cd0f7-e95a-4e74-8866-a40a65d76147 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206529, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.836359] env[62385]: DEBUG oslo_vmware.api [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206528, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.144265] env[62385]: DEBUG oslo_vmware.api [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206527, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.289401] env[62385]: DEBUG oslo_vmware.api [None req-988cd0f7-e95a-4e74-8866-a40a65d76147 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206529, 'name': PowerOffVM_Task, 'duration_secs': 0.204843} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.289762] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-988cd0f7-e95a-4e74-8866-a40a65d76147 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1122.289948] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-988cd0f7-e95a-4e74-8866-a40a65d76147 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1122.290243] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9bb4ad71-2202-4b57-9deb-3428bc9f1597 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.338905] env[62385]: DEBUG oslo_vmware.api [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206528, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.357592] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-988cd0f7-e95a-4e74-8866-a40a65d76147 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1122.357850] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-988cd0f7-e95a-4e74-8866-a40a65d76147 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Deleting contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1122.358017] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-988cd0f7-e95a-4e74-8866-a40a65d76147 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Deleting the datastore file [datastore2] 9d6f098a-0b05-43ef-96b0-9eb99ad3538c {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1122.358328] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f9a05bb7-6894-44d9-93f0-e049982191cb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.364917] env[62385]: DEBUG oslo_vmware.api [None req-988cd0f7-e95a-4e74-8866-a40a65d76147 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for the task: (returnval){ [ 1122.364917] env[62385]: value = "task-1206531" [ 1122.364917] env[62385]: _type = "Task" [ 1122.364917] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.373440] env[62385]: DEBUG oslo_vmware.api [None req-988cd0f7-e95a-4e74-8866-a40a65d76147 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206531, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.644817] env[62385]: DEBUG oslo_vmware.api [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206527, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.360496} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.645174] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/651910de-8449-4e93-bd88-46a95628369c/651910de-8449-4e93-bd88-46a95628369c.vmdk to [datastore1] c49d4366-2ea0-44f2-8463-daf62bd1c40c/c49d4366-2ea0-44f2-8463-daf62bd1c40c.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1122.645870] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49fc0a78-4e91-4c89-9272-582387e8563b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.666933] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] c49d4366-2ea0-44f2-8463-daf62bd1c40c/c49d4366-2ea0-44f2-8463-daf62bd1c40c.vmdk or device None with type streamOptimized {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1122.667155] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1f94e11b-b2bc-4839-bde6-7b20891cb0bf {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.686172] env[62385]: DEBUG oslo_vmware.api [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for the task: (returnval){ [ 1122.686172] env[62385]: value = "task-1206532" [ 1122.686172] env[62385]: _type = "Task" [ 1122.686172] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.695638] env[62385]: DEBUG oslo_vmware.api [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206532, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.837381] env[62385]: DEBUG oslo_vmware.api [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206528, 'name': ReconfigVM_Task, 'duration_secs': 1.025828} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.837670] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Reconfigured VM instance instance-0000006c to detach disk 2000 {{(pid=62385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1122.838496] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9644f7a-64f7-4206-b891-f24dbd2897c7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.860031] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317/1289e63a-5a9d-4615-b6cf-bbf5f4ae0317.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1122.860323] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6ed79c54-8ab9-49d7-b4c8-5ef7951ad64e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.881113] env[62385]: DEBUG oslo_vmware.api [None req-988cd0f7-e95a-4e74-8866-a40a65d76147 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Task: {'id': task-1206531, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.199542} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.882355] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-988cd0f7-e95a-4e74-8866-a40a65d76147 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1122.882555] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-988cd0f7-e95a-4e74-8866-a40a65d76147 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Deleted contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1122.882741] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-988cd0f7-e95a-4e74-8866-a40a65d76147 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1122.882925] env[62385]: INFO nova.compute.manager [None req-988cd0f7-e95a-4e74-8866-a40a65d76147 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1122.883190] env[62385]: DEBUG oslo.service.loopingcall [None req-988cd0f7-e95a-4e74-8866-a40a65d76147 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1122.883450] env[62385]: DEBUG oslo_vmware.api [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1122.883450] env[62385]: value = "task-1206533" [ 1122.883450] env[62385]: _type = "Task" [ 1122.883450] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.883635] env[62385]: DEBUG nova.compute.manager [-] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1122.883733] env[62385]: DEBUG nova.network.neutron [-] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1122.893851] env[62385]: DEBUG oslo_vmware.api [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206533, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.029605] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1123.029783] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Starting heal instance info cache {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1123.201100] env[62385]: DEBUG oslo_vmware.api [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206532, 'name': ReconfigVM_Task, 'duration_secs': 0.274506} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.202320] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Reconfigured VM instance instance-0000006b to attach disk [datastore1] c49d4366-2ea0-44f2-8463-daf62bd1c40c/c49d4366-2ea0-44f2-8463-daf62bd1c40c.vmdk or device None with type streamOptimized {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1123.203030] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-81100f55-c4f4-4dd7-95c1-e4304cae8c41 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.210760] env[62385]: DEBUG oslo_vmware.api [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for the task: (returnval){ [ 1123.210760] env[62385]: value = "task-1206534" [ 1123.210760] env[62385]: _type = "Task" [ 1123.210760] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.213482] env[62385]: DEBUG nova.compute.manager [req-e8c2f8a0-05d2-433f-bb58-bc5a939bce21 req-8239f145-2f19-4199-a1a2-86c83d6573c4 service nova] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Received event network-vif-deleted-ff03b686-6596-418e-b748-0a8a6799a2b5 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1123.213752] env[62385]: INFO nova.compute.manager [req-e8c2f8a0-05d2-433f-bb58-bc5a939bce21 req-8239f145-2f19-4199-a1a2-86c83d6573c4 service nova] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Neutron deleted interface ff03b686-6596-418e-b748-0a8a6799a2b5; detaching it from the instance and deleting it from the info cache [ 1123.214010] env[62385]: DEBUG nova.network.neutron [req-e8c2f8a0-05d2-433f-bb58-bc5a939bce21 req-8239f145-2f19-4199-a1a2-86c83d6573c4 service nova] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1123.231334] env[62385]: DEBUG oslo_vmware.api [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206534, 'name': Rename_Task} progress is 10%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.395477] env[62385]: DEBUG oslo_vmware.api [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206533, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.694911] env[62385]: DEBUG nova.network.neutron [-] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1123.721698] env[62385]: DEBUG oslo_vmware.api [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206534, 'name': Rename_Task, 'duration_secs': 0.135368} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.721925] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4e27a38a-2311-44f3-a98d-82282b0da95e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.723651] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1123.723878] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e5748f01-6fef-493f-a222-815c7ff933aa {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.731191] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19f5720f-9515-4a2b-a5a3-62588a91f344 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.743919] env[62385]: DEBUG oslo_vmware.api [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for the task: (returnval){ [ 1123.743919] env[62385]: value = "task-1206535" [ 1123.743919] env[62385]: _type = "Task" [ 1123.743919] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.751771] env[62385]: DEBUG oslo_vmware.api [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206535, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.757072] env[62385]: DEBUG nova.compute.manager [req-e8c2f8a0-05d2-433f-bb58-bc5a939bce21 req-8239f145-2f19-4199-a1a2-86c83d6573c4 service nova] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Detach interface failed, port_id=ff03b686-6596-418e-b748-0a8a6799a2b5, reason: Instance 9d6f098a-0b05-43ef-96b0-9eb99ad3538c could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1123.898361] env[62385]: DEBUG oslo_vmware.api [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206533, 'name': ReconfigVM_Task, 'duration_secs': 0.706802} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.898643] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Reconfigured VM instance instance-0000006c to attach disk [datastore1] 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317/1289e63a-5a9d-4615-b6cf-bbf5f4ae0317.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1123.899596] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6785de53-c3fc-4171-9407-416e052323dd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.917113] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4d3d633-31fd-477f-9ba7-f6ef440ca711 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.934102] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-002e9250-026a-405b-a83a-d1f32d63d7b7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.950764] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8503d2a4-ea69-4b73-a892-bf449b3e757f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.957058] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1123.957305] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f8629251-94ea-4b7a-bd3d-293256dd1e5c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.963123] env[62385]: DEBUG oslo_vmware.api [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1123.963123] env[62385]: value = "task-1206536" [ 1123.963123] env[62385]: _type = "Task" [ 1123.963123] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.973612] env[62385]: DEBUG oslo_vmware.api [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206536, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.197237] env[62385]: INFO nova.compute.manager [-] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Took 1.31 seconds to deallocate network for instance. [ 1124.254717] env[62385]: DEBUG oslo_vmware.api [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206535, 'name': PowerOnVM_Task, 'duration_secs': 0.455605} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.255016] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1124.360418] env[62385]: DEBUG nova.compute.manager [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1124.361436] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-038a0967-2ecf-4c68-b3bb-c47f99884cd1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.473720] env[62385]: DEBUG oslo_vmware.api [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206536, 'name': PowerOnVM_Task, 'duration_secs': 0.387435} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.474027] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1124.703428] env[62385]: DEBUG oslo_concurrency.lockutils [None req-988cd0f7-e95a-4e74-8866-a40a65d76147 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1124.703721] env[62385]: DEBUG oslo_concurrency.lockutils [None req-988cd0f7-e95a-4e74-8866-a40a65d76147 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1124.703976] env[62385]: DEBUG nova.objects.instance [None req-988cd0f7-e95a-4e74-8866-a40a65d76147 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lazy-loading 'resources' on Instance uuid 9d6f098a-0b05-43ef-96b0-9eb99ad3538c {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1124.879691] env[62385]: DEBUG oslo_concurrency.lockutils [None req-3b7b7478-146d-4af3-b29c-1eeee144bb56 tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Lock "c49d4366-2ea0-44f2-8463-daf62bd1c40c" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 19.092s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.880461] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "c49d4366-2ea0-44f2-8463-daf62bd1c40c" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 14.364s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1124.880660] env[62385]: INFO nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] During sync_power_state the instance has a pending task (spawning). Skip. [ 1124.880844] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "c49d4366-2ea0-44f2-8463-daf62bd1c40c" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1125.353336] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a72d0557-4b99-473d-8420-ef85211aea01 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.360976] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8538a61a-8f03-4dc9-9d5d-598ff64f4662 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.393238] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b92dc0fe-6800-4c5f-9ddf-52dbb2e401e6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.400814] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bb3c2b1-e76e-4b53-b820-265e6d7702f5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.416381] env[62385]: DEBUG nova.compute.provider_tree [None req-988cd0f7-e95a-4e74-8866-a40a65d76147 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1125.485350] env[62385]: INFO nova.compute.manager [None req-4c36019f-02ea-45ba-b6d0-0b8c4a535ea4 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Updating instance to original state: 'active' [ 1125.590285] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquiring lock "refresh_cache-c49d4366-2ea0-44f2-8463-daf62bd1c40c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1125.590436] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquired lock "refresh_cache-c49d4366-2ea0-44f2-8463-daf62bd1c40c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1125.590588] env[62385]: DEBUG nova.network.neutron [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Forcefully refreshing network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1125.920044] env[62385]: DEBUG nova.scheduler.client.report [None req-988cd0f7-e95a-4e74-8866-a40a65d76147 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1126.425470] env[62385]: DEBUG oslo_concurrency.lockutils [None req-988cd0f7-e95a-4e74-8866-a40a65d76147 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.722s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.446720] env[62385]: INFO nova.scheduler.client.report [None req-988cd0f7-e95a-4e74-8866-a40a65d76147 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Deleted allocations for instance 9d6f098a-0b05-43ef-96b0-9eb99ad3538c [ 1126.750631] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ece946d3-b1d6-4e25-871c-decf4db36be8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.758223] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-6354cd71-d893-4c3d-80b3-170b943d2f0e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Suspending the VM {{(pid=62385) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 1126.758485] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-dcbc6693-859b-4579-8f5a-c126b18bf73c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.765281] env[62385]: DEBUG oslo_vmware.api [None req-6354cd71-d893-4c3d-80b3-170b943d2f0e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for the task: (returnval){ [ 1126.765281] env[62385]: value = "task-1206537" [ 1126.765281] env[62385]: _type = "Task" [ 1126.765281] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.773957] env[62385]: DEBUG oslo_vmware.api [None req-6354cd71-d893-4c3d-80b3-170b943d2f0e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206537, 'name': SuspendVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.955611] env[62385]: DEBUG oslo_concurrency.lockutils [None req-988cd0f7-e95a-4e74-8866-a40a65d76147 tempest-ServerActionsTestOtherA-1278333426 tempest-ServerActionsTestOtherA-1278333426-project-member] Lock "9d6f098a-0b05-43ef-96b0-9eb99ad3538c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.198s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.959130] env[62385]: DEBUG nova.network.neutron [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Updating instance_info_cache with network_info: [{"id": "b19d5597-14de-4b08-b628-3bc8e539f77a", "address": "fa:16:3e:83:b5:c7", "network": {"id": "a561a523-c9d8-4169-b449-5c0bdef7b53b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-751953866-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d716d19f79945db98f684df0cf4b302", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb19d5597-14", "ovs_interfaceid": "b19d5597-14de-4b08-b628-3bc8e539f77a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1127.275506] env[62385]: DEBUG oslo_vmware.api [None req-6354cd71-d893-4c3d-80b3-170b943d2f0e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206537, 'name': SuspendVM_Task} progress is 66%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.463050] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Releasing lock "refresh_cache-c49d4366-2ea0-44f2-8463-daf62bd1c40c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1127.463286] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Updated the network info_cache for instance {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1127.463498] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1127.463695] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1127.463873] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1127.464021] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1127.464191] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1127.464345] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62385) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1127.571223] env[62385]: DEBUG oslo_concurrency.lockutils [None req-28797c8a-8d2f-45ec-9594-24924da20fe3 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "1289e63a-5a9d-4615-b6cf-bbf5f4ae0317" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.571517] env[62385]: DEBUG oslo_concurrency.lockutils [None req-28797c8a-8d2f-45ec-9594-24924da20fe3 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "1289e63a-5a9d-4615-b6cf-bbf5f4ae0317" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.571782] env[62385]: DEBUG oslo_concurrency.lockutils [None req-28797c8a-8d2f-45ec-9594-24924da20fe3 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "1289e63a-5a9d-4615-b6cf-bbf5f4ae0317-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.572045] env[62385]: DEBUG oslo_concurrency.lockutils [None req-28797c8a-8d2f-45ec-9594-24924da20fe3 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "1289e63a-5a9d-4615-b6cf-bbf5f4ae0317-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.572167] env[62385]: DEBUG oslo_concurrency.lockutils [None req-28797c8a-8d2f-45ec-9594-24924da20fe3 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "1289e63a-5a9d-4615-b6cf-bbf5f4ae0317-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1127.574452] env[62385]: INFO nova.compute.manager [None req-28797c8a-8d2f-45ec-9594-24924da20fe3 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Terminating instance [ 1127.576296] env[62385]: DEBUG nova.compute.manager [None req-28797c8a-8d2f-45ec-9594-24924da20fe3 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1127.576500] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-28797c8a-8d2f-45ec-9594-24924da20fe3 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1127.577360] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b23e926-0a16-448d-8193-e4e46937d228 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.584908] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-28797c8a-8d2f-45ec-9594-24924da20fe3 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1127.585160] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5f384ae5-76a5-4448-81f3-31a83e736622 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.592516] env[62385]: DEBUG oslo_vmware.api [None req-28797c8a-8d2f-45ec-9594-24924da20fe3 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1127.592516] env[62385]: value = "task-1206538" [ 1127.592516] env[62385]: _type = "Task" [ 1127.592516] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.600781] env[62385]: DEBUG oslo_vmware.api [None req-28797c8a-8d2f-45ec-9594-24924da20fe3 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206538, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.776560] env[62385]: DEBUG oslo_vmware.api [None req-6354cd71-d893-4c3d-80b3-170b943d2f0e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206537, 'name': SuspendVM_Task, 'duration_secs': 0.62054} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.776837] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-6354cd71-d893-4c3d-80b3-170b943d2f0e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Suspended the VM {{(pid=62385) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 1127.777018] env[62385]: DEBUG nova.compute.manager [None req-6354cd71-d893-4c3d-80b3-170b943d2f0e tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1127.777855] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af89b1c0-8888-46d7-b2bd-6915fa693fee {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.029484] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1128.029770] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager.update_available_resource {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1128.105815] env[62385]: DEBUG oslo_vmware.api [None req-28797c8a-8d2f-45ec-9594-24924da20fe3 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206538, 'name': PowerOffVM_Task, 'duration_secs': 0.25447} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.105815] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-28797c8a-8d2f-45ec-9594-24924da20fe3 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1128.105815] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-28797c8a-8d2f-45ec-9594-24924da20fe3 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1128.105815] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d9fe7bd4-374d-42a1-99f5-922b1a4e692c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.183986] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-28797c8a-8d2f-45ec-9594-24924da20fe3 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1128.184306] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-28797c8a-8d2f-45ec-9594-24924da20fe3 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1128.184439] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-28797c8a-8d2f-45ec-9594-24924da20fe3 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Deleting the datastore file [datastore1] 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1128.184662] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8627c5ac-f90e-4d37-909e-541d01fca15b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.191601] env[62385]: DEBUG oslo_vmware.api [None req-28797c8a-8d2f-45ec-9594-24924da20fe3 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1128.191601] env[62385]: value = "task-1206540" [ 1128.191601] env[62385]: _type = "Task" [ 1128.191601] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.199182] env[62385]: DEBUG oslo_vmware.api [None req-28797c8a-8d2f-45ec-9594-24924da20fe3 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206540, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.532803] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1128.533065] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1128.533280] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1128.533406] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62385) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1128.534317] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88f04532-e495-4cdb-9943-4fffa15bd836 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.543172] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2c36882-50af-4c05-a855-11421db5fd0a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.560623] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-028d7803-84ef-4d4d-91fb-48fb01cafc2b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.567573] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4796a35-7437-4abc-b9cb-6a599eea104d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.597715] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180645MB free_disk=155GB free_vcpus=48 pci_devices=None {{(pid=62385) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1128.597885] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1128.601718] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1128.702268] env[62385]: DEBUG oslo_vmware.api [None req-28797c8a-8d2f-45ec-9594-24924da20fe3 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206540, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.251868} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.703524] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-28797c8a-8d2f-45ec-9594-24924da20fe3 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1128.703819] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-28797c8a-8d2f-45ec-9594-24924da20fe3 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1128.703941] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-28797c8a-8d2f-45ec-9594-24924da20fe3 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1128.704124] env[62385]: INFO nova.compute.manager [None req-28797c8a-8d2f-45ec-9594-24924da20fe3 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1128.704371] env[62385]: DEBUG oslo.service.loopingcall [None req-28797c8a-8d2f-45ec-9594-24924da20fe3 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1128.706342] env[62385]: DEBUG nova.compute.manager [-] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1128.706678] env[62385]: DEBUG nova.network.neutron [-] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1129.209790] env[62385]: INFO nova.compute.manager [None req-00e847f2-e3f9-4807-9e16-bde30b6ba21a tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Resuming [ 1129.210461] env[62385]: DEBUG nova.objects.instance [None req-00e847f2-e3f9-4807-9e16-bde30b6ba21a tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Lazy-loading 'flavor' on Instance uuid c49d4366-2ea0-44f2-8463-daf62bd1c40c {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1129.277534] env[62385]: DEBUG nova.compute.manager [req-5d28babc-005f-49ce-a9ff-2278221535e5 req-3b5e0a6c-78fb-4e6e-84d8-6629cb204207 service nova] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Received event network-vif-deleted-c6ede9da-b209-4fae-b75a-c2f1ede4dcaf {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1129.277735] env[62385]: INFO nova.compute.manager [req-5d28babc-005f-49ce-a9ff-2278221535e5 req-3b5e0a6c-78fb-4e6e-84d8-6629cb204207 service nova] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Neutron deleted interface c6ede9da-b209-4fae-b75a-c2f1ede4dcaf; detaching it from the instance and deleting it from the info cache [ 1129.277915] env[62385]: DEBUG nova.network.neutron [req-5d28babc-005f-49ce-a9ff-2278221535e5 req-3b5e0a6c-78fb-4e6e-84d8-6629cb204207 service nova] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1129.626625] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 4fe375be-ba60-4601-bd89-610e12742df3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1129.626811] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance c49d4366-2ea0-44f2-8463-daf62bd1c40c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1129.626940] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1129.627143] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=62385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1129.627283] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=62385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1129.657388] env[62385]: DEBUG nova.network.neutron [-] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1129.686053] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23f2eefa-5aa9-4357-a788-525c714b3592 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.693784] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b86f1892-9f37-4eb2-a6a0-62c2a864c3b8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.725484] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1884be71-a866-4da6-aaea-3d918ab9ad45 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.735606] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d3004df-0e3e-48c6-9313-ff26aa0fdbb9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.748024] env[62385]: DEBUG nova.compute.provider_tree [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1129.781769] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-94777065-95a8-4ca6-b31f-1e7dd93c19f2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.791290] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a28e92e-0b78-4fb5-96f6-27219897bc3f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.813732] env[62385]: DEBUG nova.compute.manager [req-5d28babc-005f-49ce-a9ff-2278221535e5 req-3b5e0a6c-78fb-4e6e-84d8-6629cb204207 service nova] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Detach interface failed, port_id=c6ede9da-b209-4fae-b75a-c2f1ede4dcaf, reason: Instance 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317 could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1130.160788] env[62385]: INFO nova.compute.manager [-] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Took 1.45 seconds to deallocate network for instance. [ 1130.230774] env[62385]: DEBUG oslo_concurrency.lockutils [None req-00e847f2-e3f9-4807-9e16-bde30b6ba21a tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Acquiring lock "refresh_cache-c49d4366-2ea0-44f2-8463-daf62bd1c40c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1130.231261] env[62385]: DEBUG oslo_concurrency.lockutils [None req-00e847f2-e3f9-4807-9e16-bde30b6ba21a tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Acquired lock "refresh_cache-c49d4366-2ea0-44f2-8463-daf62bd1c40c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1130.231261] env[62385]: DEBUG nova.network.neutron [None req-00e847f2-e3f9-4807-9e16-bde30b6ba21a tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1130.249855] env[62385]: DEBUG nova.scheduler.client.report [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1130.667830] env[62385]: DEBUG oslo_concurrency.lockutils [None req-28797c8a-8d2f-45ec-9594-24924da20fe3 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1130.758434] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62385) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1130.758645] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.161s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1130.758978] env[62385]: DEBUG oslo_concurrency.lockutils [None req-28797c8a-8d2f-45ec-9594-24924da20fe3 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.091s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1130.759231] env[62385]: DEBUG nova.objects.instance [None req-28797c8a-8d2f-45ec-9594-24924da20fe3 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lazy-loading 'resources' on Instance uuid 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1130.929897] env[62385]: DEBUG nova.network.neutron [None req-00e847f2-e3f9-4807-9e16-bde30b6ba21a tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Updating instance_info_cache with network_info: [{"id": "b19d5597-14de-4b08-b628-3bc8e539f77a", "address": "fa:16:3e:83:b5:c7", "network": {"id": "a561a523-c9d8-4169-b449-5c0bdef7b53b", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-751953866-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d716d19f79945db98f684df0cf4b302", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb19d5597-14", "ovs_interfaceid": "b19d5597-14de-4b08-b628-3bc8e539f77a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1131.336137] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17ef2556-b791-4c79-9f38-8c62ab203caa {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.344270] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b302e9b5-726a-4675-9555-6a37b24dabc0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.375877] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37698e38-b7f2-44c4-9b90-0d29c400f2bf {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.383572] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8436f62d-33cf-4df2-bc35-08dd6f5520f6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.397283] env[62385]: DEBUG nova.compute.provider_tree [None req-28797c8a-8d2f-45ec-9594-24924da20fe3 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1131.432492] env[62385]: DEBUG oslo_concurrency.lockutils [None req-00e847f2-e3f9-4807-9e16-bde30b6ba21a tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Releasing lock "refresh_cache-c49d4366-2ea0-44f2-8463-daf62bd1c40c" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1131.433726] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a88be21e-e15c-4b8d-9ef1-a140636b9418 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.440681] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-00e847f2-e3f9-4807-9e16-bde30b6ba21a tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Resuming the VM {{(pid=62385) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1184}} [ 1131.441372] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b2691599-a052-42e5-9382-d207f7a14e8e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.448623] env[62385]: DEBUG oslo_vmware.api [None req-00e847f2-e3f9-4807-9e16-bde30b6ba21a tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for the task: (returnval){ [ 1131.448623] env[62385]: value = "task-1206541" [ 1131.448623] env[62385]: _type = "Task" [ 1131.448623] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.458368] env[62385]: DEBUG oslo_vmware.api [None req-00e847f2-e3f9-4807-9e16-bde30b6ba21a tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206541, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.900188] env[62385]: DEBUG nova.scheduler.client.report [None req-28797c8a-8d2f-45ec-9594-24924da20fe3 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1131.963525] env[62385]: DEBUG oslo_vmware.api [None req-00e847f2-e3f9-4807-9e16-bde30b6ba21a tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206541, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.405245] env[62385]: DEBUG oslo_concurrency.lockutils [None req-28797c8a-8d2f-45ec-9594-24924da20fe3 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.646s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1132.430139] env[62385]: INFO nova.scheduler.client.report [None req-28797c8a-8d2f-45ec-9594-24924da20fe3 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Deleted allocations for instance 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317 [ 1132.465297] env[62385]: DEBUG oslo_vmware.api [None req-00e847f2-e3f9-4807-9e16-bde30b6ba21a tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206541, 'name': PowerOnVM_Task, 'duration_secs': 0.522871} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.467493] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-00e847f2-e3f9-4807-9e16-bde30b6ba21a tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Resumed the VM {{(pid=62385) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1189}} [ 1132.467718] env[62385]: DEBUG nova.compute.manager [None req-00e847f2-e3f9-4807-9e16-bde30b6ba21a tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1132.468550] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cada881-d655-45c7-8027-5d28e836a7db {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.941782] env[62385]: DEBUG oslo_concurrency.lockutils [None req-28797c8a-8d2f-45ec-9594-24924da20fe3 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "1289e63a-5a9d-4615-b6cf-bbf5f4ae0317" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.370s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.804164] env[62385]: DEBUG oslo_concurrency.lockutils [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "28afc06e-f6e4-4044-8bfe-c93b1d9c1388" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1133.804493] env[62385]: DEBUG oslo_concurrency.lockutils [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "28afc06e-f6e4-4044-8bfe-c93b1d9c1388" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1134.306574] env[62385]: DEBUG nova.compute.manager [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1134.829834] env[62385]: DEBUG oslo_concurrency.lockutils [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1134.830122] env[62385]: DEBUG oslo_concurrency.lockutils [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1134.831876] env[62385]: INFO nova.compute.claims [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1135.887921] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa8307ab-5104-4b8f-9e65-60307ecbd8ec {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.895922] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfc9d7a9-2352-43e3-a218-9a2fdffea27b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.926015] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cc0ca6a-2db1-4cb6-a396-80f9a272071a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.932541] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecb68df2-6a41-4b57-bd12-5471c272cb4e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.944896] env[62385]: DEBUG nova.compute.provider_tree [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1136.447892] env[62385]: DEBUG nova.scheduler.client.report [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1136.953163] env[62385]: DEBUG oslo_concurrency.lockutils [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.123s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1136.953706] env[62385]: DEBUG nova.compute.manager [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1137.459028] env[62385]: DEBUG nova.compute.utils [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1137.459682] env[62385]: DEBUG nova.compute.manager [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1137.459863] env[62385]: DEBUG nova.network.neutron [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1137.496414] env[62385]: DEBUG nova.policy [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7e04c6591fe043d686140eb2f5ce0be2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd45b5346eab44707b4d81eaaf12b0935', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 1137.631871] env[62385]: DEBUG oslo_concurrency.lockutils [None req-78000643-aa06-4cf3-a527-358db7fc56fe tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Acquiring lock "c49d4366-2ea0-44f2-8463-daf62bd1c40c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.632181] env[62385]: DEBUG oslo_concurrency.lockutils [None req-78000643-aa06-4cf3-a527-358db7fc56fe tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Lock "c49d4366-2ea0-44f2-8463-daf62bd1c40c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.632413] env[62385]: DEBUG oslo_concurrency.lockutils [None req-78000643-aa06-4cf3-a527-358db7fc56fe tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Acquiring lock "c49d4366-2ea0-44f2-8463-daf62bd1c40c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.632616] env[62385]: DEBUG oslo_concurrency.lockutils [None req-78000643-aa06-4cf3-a527-358db7fc56fe tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Lock "c49d4366-2ea0-44f2-8463-daf62bd1c40c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.632791] env[62385]: DEBUG oslo_concurrency.lockutils [None req-78000643-aa06-4cf3-a527-358db7fc56fe tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Lock "c49d4366-2ea0-44f2-8463-daf62bd1c40c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.634935] env[62385]: INFO nova.compute.manager [None req-78000643-aa06-4cf3-a527-358db7fc56fe tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Terminating instance [ 1137.636729] env[62385]: DEBUG nova.compute.manager [None req-78000643-aa06-4cf3-a527-358db7fc56fe tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1137.637614] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-78000643-aa06-4cf3-a527-358db7fc56fe tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1137.637694] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-522aa2b9-3987-4be7-98a6-66fda9a29a32 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.645407] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-78000643-aa06-4cf3-a527-358db7fc56fe tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1137.645624] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6c750d8d-594e-4e8f-b44d-38bbb7c50d33 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.652132] env[62385]: DEBUG oslo_vmware.api [None req-78000643-aa06-4cf3-a527-358db7fc56fe tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for the task: (returnval){ [ 1137.652132] env[62385]: value = "task-1206542" [ 1137.652132] env[62385]: _type = "Task" [ 1137.652132] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.659681] env[62385]: DEBUG oslo_vmware.api [None req-78000643-aa06-4cf3-a527-358db7fc56fe tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206542, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.769091] env[62385]: DEBUG nova.network.neutron [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Successfully created port: 825ca180-3836-41d5-8d91-f55bc2ec9853 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1137.963764] env[62385]: DEBUG nova.compute.manager [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1138.161712] env[62385]: DEBUG oslo_vmware.api [None req-78000643-aa06-4cf3-a527-358db7fc56fe tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206542, 'name': PowerOffVM_Task, 'duration_secs': 0.180242} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.162048] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-78000643-aa06-4cf3-a527-358db7fc56fe tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1138.162232] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-78000643-aa06-4cf3-a527-358db7fc56fe tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1138.162480] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-64743415-9c9f-4b54-aed3-4ca504bf8a4e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.221701] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-78000643-aa06-4cf3-a527-358db7fc56fe tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1138.221981] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-78000643-aa06-4cf3-a527-358db7fc56fe tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1138.222199] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-78000643-aa06-4cf3-a527-358db7fc56fe tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Deleting the datastore file [datastore1] c49d4366-2ea0-44f2-8463-daf62bd1c40c {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1138.222472] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b94f1c75-71cd-4d18-9efb-8f27f468f2c3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.229895] env[62385]: DEBUG oslo_vmware.api [None req-78000643-aa06-4cf3-a527-358db7fc56fe tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for the task: (returnval){ [ 1138.229895] env[62385]: value = "task-1206544" [ 1138.229895] env[62385]: _type = "Task" [ 1138.229895] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.237366] env[62385]: DEBUG oslo_vmware.api [None req-78000643-aa06-4cf3-a527-358db7fc56fe tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206544, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.739696] env[62385]: DEBUG oslo_vmware.api [None req-78000643-aa06-4cf3-a527-358db7fc56fe tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Task: {'id': task-1206544, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140698} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.739957] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-78000643-aa06-4cf3-a527-358db7fc56fe tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1138.740192] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-78000643-aa06-4cf3-a527-358db7fc56fe tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1138.740371] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-78000643-aa06-4cf3-a527-358db7fc56fe tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1138.740545] env[62385]: INFO nova.compute.manager [None req-78000643-aa06-4cf3-a527-358db7fc56fe tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1138.740786] env[62385]: DEBUG oslo.service.loopingcall [None req-78000643-aa06-4cf3-a527-358db7fc56fe tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1138.740979] env[62385]: DEBUG nova.compute.manager [-] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1138.741102] env[62385]: DEBUG nova.network.neutron [-] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1138.972664] env[62385]: DEBUG nova.compute.manager [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1138.997039] env[62385]: DEBUG nova.virt.hardware [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1138.997319] env[62385]: DEBUG nova.virt.hardware [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1138.997474] env[62385]: DEBUG nova.virt.hardware [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1138.997634] env[62385]: DEBUG nova.virt.hardware [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1138.997782] env[62385]: DEBUG nova.virt.hardware [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1138.998041] env[62385]: DEBUG nova.virt.hardware [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1138.998288] env[62385]: DEBUG nova.virt.hardware [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1138.998577] env[62385]: DEBUG nova.virt.hardware [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1138.998655] env[62385]: DEBUG nova.virt.hardware [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1138.998784] env[62385]: DEBUG nova.virt.hardware [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1138.998965] env[62385]: DEBUG nova.virt.hardware [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1138.999845] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35a585a0-3fda-40dd-946b-e6cc14fc2860 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.008167] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15cf23a5-ca46-4186-babc-a9421b23243f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.013975] env[62385]: DEBUG nova.compute.manager [req-890e012a-bdb5-4e24-88e1-de3bda88532d req-4dee6297-4ab5-43d4-bcf8-e3e13faaefee service nova] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Received event network-vif-deleted-b19d5597-14de-4b08-b628-3bc8e539f77a {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1139.014174] env[62385]: INFO nova.compute.manager [req-890e012a-bdb5-4e24-88e1-de3bda88532d req-4dee6297-4ab5-43d4-bcf8-e3e13faaefee service nova] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Neutron deleted interface b19d5597-14de-4b08-b628-3bc8e539f77a; detaching it from the instance and deleting it from the info cache [ 1139.014330] env[62385]: DEBUG nova.network.neutron [req-890e012a-bdb5-4e24-88e1-de3bda88532d req-4dee6297-4ab5-43d4-bcf8-e3e13faaefee service nova] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1139.137245] env[62385]: DEBUG nova.compute.manager [req-c929128d-364c-4325-bcb6-565db5b82b54 req-b81826da-732e-4dd1-9ffc-26483499bbb9 service nova] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Received event network-vif-plugged-825ca180-3836-41d5-8d91-f55bc2ec9853 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1139.137470] env[62385]: DEBUG oslo_concurrency.lockutils [req-c929128d-364c-4325-bcb6-565db5b82b54 req-b81826da-732e-4dd1-9ffc-26483499bbb9 service nova] Acquiring lock "28afc06e-f6e4-4044-8bfe-c93b1d9c1388-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1139.137696] env[62385]: DEBUG oslo_concurrency.lockutils [req-c929128d-364c-4325-bcb6-565db5b82b54 req-b81826da-732e-4dd1-9ffc-26483499bbb9 service nova] Lock "28afc06e-f6e4-4044-8bfe-c93b1d9c1388-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1139.137875] env[62385]: DEBUG oslo_concurrency.lockutils [req-c929128d-364c-4325-bcb6-565db5b82b54 req-b81826da-732e-4dd1-9ffc-26483499bbb9 service nova] Lock "28afc06e-f6e4-4044-8bfe-c93b1d9c1388-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1139.138084] env[62385]: DEBUG nova.compute.manager [req-c929128d-364c-4325-bcb6-565db5b82b54 req-b81826da-732e-4dd1-9ffc-26483499bbb9 service nova] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] No waiting events found dispatching network-vif-plugged-825ca180-3836-41d5-8d91-f55bc2ec9853 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1139.138475] env[62385]: WARNING nova.compute.manager [req-c929128d-364c-4325-bcb6-565db5b82b54 req-b81826da-732e-4dd1-9ffc-26483499bbb9 service nova] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Received unexpected event network-vif-plugged-825ca180-3836-41d5-8d91-f55bc2ec9853 for instance with vm_state building and task_state spawning. [ 1139.219917] env[62385]: DEBUG nova.network.neutron [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Successfully updated port: 825ca180-3836-41d5-8d91-f55bc2ec9853 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1139.464446] env[62385]: DEBUG nova.network.neutron [-] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1139.516794] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d329acfc-14bc-40c7-ae93-1fb670a6a3f0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.526553] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-306d9691-67f4-403f-86f6-75ba7d001cb7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.549317] env[62385]: DEBUG nova.compute.manager [req-890e012a-bdb5-4e24-88e1-de3bda88532d req-4dee6297-4ab5-43d4-bcf8-e3e13faaefee service nova] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Detach interface failed, port_id=b19d5597-14de-4b08-b628-3bc8e539f77a, reason: Instance c49d4366-2ea0-44f2-8463-daf62bd1c40c could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1139.723591] env[62385]: DEBUG oslo_concurrency.lockutils [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "refresh_cache-28afc06e-f6e4-4044-8bfe-c93b1d9c1388" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1139.723591] env[62385]: DEBUG oslo_concurrency.lockutils [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquired lock "refresh_cache-28afc06e-f6e4-4044-8bfe-c93b1d9c1388" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1139.723591] env[62385]: DEBUG nova.network.neutron [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1139.966967] env[62385]: INFO nova.compute.manager [-] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Took 1.23 seconds to deallocate network for instance. [ 1140.254224] env[62385]: DEBUG nova.network.neutron [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1140.407497] env[62385]: DEBUG nova.network.neutron [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Updating instance_info_cache with network_info: [{"id": "825ca180-3836-41d5-8d91-f55bc2ec9853", "address": "fa:16:3e:e7:99:34", "network": {"id": "06017161-984c-4657-9639-4e6bdb6d4c09", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1072882119-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d45b5346eab44707b4d81eaaf12b0935", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39cd75b0-9ec7-48ed-b57f-34da0c573a60", "external-id": "nsx-vlan-transportzone-751", "segmentation_id": 751, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap825ca180-38", "ovs_interfaceid": "825ca180-3836-41d5-8d91-f55bc2ec9853", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1140.473298] env[62385]: DEBUG oslo_concurrency.lockutils [None req-78000643-aa06-4cf3-a527-358db7fc56fe tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.473702] env[62385]: DEBUG oslo_concurrency.lockutils [None req-78000643-aa06-4cf3-a527-358db7fc56fe tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.473797] env[62385]: DEBUG nova.objects.instance [None req-78000643-aa06-4cf3-a527-358db7fc56fe tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Lazy-loading 'resources' on Instance uuid c49d4366-2ea0-44f2-8463-daf62bd1c40c {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1140.911343] env[62385]: DEBUG oslo_concurrency.lockutils [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Releasing lock "refresh_cache-28afc06e-f6e4-4044-8bfe-c93b1d9c1388" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1140.911673] env[62385]: DEBUG nova.compute.manager [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Instance network_info: |[{"id": "825ca180-3836-41d5-8d91-f55bc2ec9853", "address": "fa:16:3e:e7:99:34", "network": {"id": "06017161-984c-4657-9639-4e6bdb6d4c09", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1072882119-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d45b5346eab44707b4d81eaaf12b0935", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39cd75b0-9ec7-48ed-b57f-34da0c573a60", "external-id": "nsx-vlan-transportzone-751", "segmentation_id": 751, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap825ca180-38", "ovs_interfaceid": "825ca180-3836-41d5-8d91-f55bc2ec9853", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1140.912179] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e7:99:34', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '39cd75b0-9ec7-48ed-b57f-34da0c573a60', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '825ca180-3836-41d5-8d91-f55bc2ec9853', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1140.919539] env[62385]: DEBUG oslo.service.loopingcall [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1140.919779] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1140.920017] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c4fd79b1-4377-48df-972e-59f14e7623cb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.939446] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1140.939446] env[62385]: value = "task-1206545" [ 1140.939446] env[62385]: _type = "Task" [ 1140.939446] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.946778] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206545, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.031348] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bab06576-73bb-441c-be5c-73e149af0082 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.038644] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19e8b19f-d45b-4b67-b524-66075c672d80 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.067719] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bfd7a90-3f25-47c6-a95d-fc7705d52b5d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.075031] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff85a2b4-e424-434a-8c3f-482a4b3b065c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.089594] env[62385]: DEBUG nova.compute.provider_tree [None req-78000643-aa06-4cf3-a527-358db7fc56fe tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1141.162744] env[62385]: DEBUG nova.compute.manager [req-87995f98-8c12-46b5-b3f2-d02b8a744be0 req-9638ed83-31de-4442-be75-adbd1cf2ff14 service nova] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Received event network-changed-825ca180-3836-41d5-8d91-f55bc2ec9853 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1141.162946] env[62385]: DEBUG nova.compute.manager [req-87995f98-8c12-46b5-b3f2-d02b8a744be0 req-9638ed83-31de-4442-be75-adbd1cf2ff14 service nova] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Refreshing instance network info cache due to event network-changed-825ca180-3836-41d5-8d91-f55bc2ec9853. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1141.163205] env[62385]: DEBUG oslo_concurrency.lockutils [req-87995f98-8c12-46b5-b3f2-d02b8a744be0 req-9638ed83-31de-4442-be75-adbd1cf2ff14 service nova] Acquiring lock "refresh_cache-28afc06e-f6e4-4044-8bfe-c93b1d9c1388" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1141.163373] env[62385]: DEBUG oslo_concurrency.lockutils [req-87995f98-8c12-46b5-b3f2-d02b8a744be0 req-9638ed83-31de-4442-be75-adbd1cf2ff14 service nova] Acquired lock "refresh_cache-28afc06e-f6e4-4044-8bfe-c93b1d9c1388" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1141.163540] env[62385]: DEBUG nova.network.neutron [req-87995f98-8c12-46b5-b3f2-d02b8a744be0 req-9638ed83-31de-4442-be75-adbd1cf2ff14 service nova] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Refreshing network info cache for port 825ca180-3836-41d5-8d91-f55bc2ec9853 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1141.449580] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206545, 'name': CreateVM_Task, 'duration_secs': 0.341097} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.449972] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1141.450632] env[62385]: DEBUG oslo_concurrency.lockutils [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1141.450829] env[62385]: DEBUG oslo_concurrency.lockutils [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1141.451227] env[62385]: DEBUG oslo_concurrency.lockutils [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1141.451507] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7fdfab75-bc39-47a7-9d2f-811c162fcf6d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.455888] env[62385]: DEBUG oslo_vmware.api [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1141.455888] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]524c7101-39bf-022e-52a1-55dada303ad3" [ 1141.455888] env[62385]: _type = "Task" [ 1141.455888] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.462827] env[62385]: DEBUG oslo_vmware.api [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]524c7101-39bf-022e-52a1-55dada303ad3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.593305] env[62385]: DEBUG nova.scheduler.client.report [None req-78000643-aa06-4cf3-a527-358db7fc56fe tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1141.844579] env[62385]: DEBUG nova.network.neutron [req-87995f98-8c12-46b5-b3f2-d02b8a744be0 req-9638ed83-31de-4442-be75-adbd1cf2ff14 service nova] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Updated VIF entry in instance network info cache for port 825ca180-3836-41d5-8d91-f55bc2ec9853. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1141.844945] env[62385]: DEBUG nova.network.neutron [req-87995f98-8c12-46b5-b3f2-d02b8a744be0 req-9638ed83-31de-4442-be75-adbd1cf2ff14 service nova] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Updating instance_info_cache with network_info: [{"id": "825ca180-3836-41d5-8d91-f55bc2ec9853", "address": "fa:16:3e:e7:99:34", "network": {"id": "06017161-984c-4657-9639-4e6bdb6d4c09", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1072882119-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d45b5346eab44707b4d81eaaf12b0935", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39cd75b0-9ec7-48ed-b57f-34da0c573a60", "external-id": "nsx-vlan-transportzone-751", "segmentation_id": 751, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap825ca180-38", "ovs_interfaceid": "825ca180-3836-41d5-8d91-f55bc2ec9853", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1141.966818] env[62385]: DEBUG oslo_vmware.api [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]524c7101-39bf-022e-52a1-55dada303ad3, 'name': SearchDatastore_Task, 'duration_secs': 0.01102} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.967135] env[62385]: DEBUG oslo_concurrency.lockutils [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1141.967383] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1141.967618] env[62385]: DEBUG oslo_concurrency.lockutils [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1141.967794] env[62385]: DEBUG oslo_concurrency.lockutils [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1141.968062] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1141.968330] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-63b3a5b8-8e73-4d19-b42c-c7cef24338b4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.975798] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1141.975975] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1141.976652] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7e4d37f8-e97f-4aea-a1fe-43a0e3fcf3fb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.981370] env[62385]: DEBUG oslo_vmware.api [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1141.981370] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52753469-7383-d97f-a2a6-5962e5a503d2" [ 1141.981370] env[62385]: _type = "Task" [ 1141.981370] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.988624] env[62385]: DEBUG oslo_vmware.api [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52753469-7383-d97f-a2a6-5962e5a503d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.098765] env[62385]: DEBUG oslo_concurrency.lockutils [None req-78000643-aa06-4cf3-a527-358db7fc56fe tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.625s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.120687] env[62385]: INFO nova.scheduler.client.report [None req-78000643-aa06-4cf3-a527-358db7fc56fe tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Deleted allocations for instance c49d4366-2ea0-44f2-8463-daf62bd1c40c [ 1142.347457] env[62385]: DEBUG oslo_concurrency.lockutils [req-87995f98-8c12-46b5-b3f2-d02b8a744be0 req-9638ed83-31de-4442-be75-adbd1cf2ff14 service nova] Releasing lock "refresh_cache-28afc06e-f6e4-4044-8bfe-c93b1d9c1388" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1142.492395] env[62385]: DEBUG oslo_vmware.api [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52753469-7383-d97f-a2a6-5962e5a503d2, 'name': SearchDatastore_Task, 'duration_secs': 0.008232} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.493104] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91b6d5ab-f4b6-4ec7-a4e8-68c6e8de7786 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.498231] env[62385]: DEBUG oslo_vmware.api [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1142.498231] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52b7527c-e5be-72eb-9199-f90917fe3d58" [ 1142.498231] env[62385]: _type = "Task" [ 1142.498231] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.505303] env[62385]: DEBUG oslo_vmware.api [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52b7527c-e5be-72eb-9199-f90917fe3d58, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.628689] env[62385]: DEBUG oslo_concurrency.lockutils [None req-78000643-aa06-4cf3-a527-358db7fc56fe tempest-ServersNegativeTestJSON-792721539 tempest-ServersNegativeTestJSON-792721539-project-member] Lock "c49d4366-2ea0-44f2-8463-daf62bd1c40c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.996s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1143.011446] env[62385]: DEBUG oslo_vmware.api [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52b7527c-e5be-72eb-9199-f90917fe3d58, 'name': SearchDatastore_Task, 'duration_secs': 0.013636} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.011446] env[62385]: DEBUG oslo_concurrency.lockutils [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1143.011446] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 28afc06e-f6e4-4044-8bfe-c93b1d9c1388/28afc06e-f6e4-4044-8bfe-c93b1d9c1388.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1143.011446] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c6fe4946-59e4-4082-bb63-0c27610161e9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.018345] env[62385]: DEBUG oslo_vmware.api [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1143.018345] env[62385]: value = "task-1206546" [ 1143.018345] env[62385]: _type = "Task" [ 1143.018345] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.026226] env[62385]: DEBUG oslo_vmware.api [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206546, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.528726] env[62385]: DEBUG oslo_vmware.api [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206546, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.441802} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.529138] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore2] 28afc06e-f6e4-4044-8bfe-c93b1d9c1388/28afc06e-f6e4-4044-8bfe-c93b1d9c1388.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1143.529238] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1143.529495] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5084909e-329b-457d-90fb-13b3e8094dcc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.535742] env[62385]: DEBUG oslo_vmware.api [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1143.535742] env[62385]: value = "task-1206547" [ 1143.535742] env[62385]: _type = "Task" [ 1143.535742] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.543119] env[62385]: DEBUG oslo_vmware.api [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206547, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.045187] env[62385]: DEBUG oslo_vmware.api [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206547, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080407} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.045462] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1144.046247] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa730a37-bd0a-42ab-8450-7eeb9062d7c6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.067921] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Reconfiguring VM instance instance-00000072 to attach disk [datastore2] 28afc06e-f6e4-4044-8bfe-c93b1d9c1388/28afc06e-f6e4-4044-8bfe-c93b1d9c1388.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1144.068220] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-297faf57-2976-4e69-bb0e-cf4147287398 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.088327] env[62385]: DEBUG oslo_vmware.api [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1144.088327] env[62385]: value = "task-1206548" [ 1144.088327] env[62385]: _type = "Task" [ 1144.088327] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.096056] env[62385]: DEBUG oslo_vmware.api [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206548, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.598450] env[62385]: DEBUG oslo_vmware.api [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206548, 'name': ReconfigVM_Task, 'duration_secs': 0.263527} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.598754] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Reconfigured VM instance instance-00000072 to attach disk [datastore2] 28afc06e-f6e4-4044-8bfe-c93b1d9c1388/28afc06e-f6e4-4044-8bfe-c93b1d9c1388.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1144.599578] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3707ca0a-4d72-4ecc-bfba-c249437b3615 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.605569] env[62385]: DEBUG oslo_vmware.api [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1144.605569] env[62385]: value = "task-1206549" [ 1144.605569] env[62385]: _type = "Task" [ 1144.605569] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.613018] env[62385]: DEBUG oslo_vmware.api [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206549, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.115866] env[62385]: DEBUG oslo_vmware.api [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206549, 'name': Rename_Task, 'duration_secs': 0.143563} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.116160] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1145.116419] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a8bc5e4f-c420-4f0b-afd5-648d794da9f4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.122955] env[62385]: DEBUG oslo_vmware.api [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1145.122955] env[62385]: value = "task-1206550" [ 1145.122955] env[62385]: _type = "Task" [ 1145.122955] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.130240] env[62385]: DEBUG oslo_vmware.api [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206550, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.632984] env[62385]: DEBUG oslo_vmware.api [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206550, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.133016] env[62385]: DEBUG oslo_vmware.api [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206550, 'name': PowerOnVM_Task, 'duration_secs': 0.689451} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.133328] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1146.133537] env[62385]: INFO nova.compute.manager [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Took 7.16 seconds to spawn the instance on the hypervisor. [ 1146.133720] env[62385]: DEBUG nova.compute.manager [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1146.134557] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edf786bc-24d1-4536-8430-23382fd063ee {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.654584] env[62385]: INFO nova.compute.manager [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Took 11.84 seconds to build instance. [ 1146.940733] env[62385]: DEBUG nova.compute.manager [req-c59eee15-b5c0-4de8-b332-c7cf96bf15db req-bdf6da5d-b99c-473c-bdad-f7a9ed286cce service nova] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Received event network-changed-825ca180-3836-41d5-8d91-f55bc2ec9853 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1146.940925] env[62385]: DEBUG nova.compute.manager [req-c59eee15-b5c0-4de8-b332-c7cf96bf15db req-bdf6da5d-b99c-473c-bdad-f7a9ed286cce service nova] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Refreshing instance network info cache due to event network-changed-825ca180-3836-41d5-8d91-f55bc2ec9853. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1146.941102] env[62385]: DEBUG oslo_concurrency.lockutils [req-c59eee15-b5c0-4de8-b332-c7cf96bf15db req-bdf6da5d-b99c-473c-bdad-f7a9ed286cce service nova] Acquiring lock "refresh_cache-28afc06e-f6e4-4044-8bfe-c93b1d9c1388" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1146.941310] env[62385]: DEBUG oslo_concurrency.lockutils [req-c59eee15-b5c0-4de8-b332-c7cf96bf15db req-bdf6da5d-b99c-473c-bdad-f7a9ed286cce service nova] Acquired lock "refresh_cache-28afc06e-f6e4-4044-8bfe-c93b1d9c1388" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1146.941420] env[62385]: DEBUG nova.network.neutron [req-c59eee15-b5c0-4de8-b332-c7cf96bf15db req-bdf6da5d-b99c-473c-bdad-f7a9ed286cce service nova] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Refreshing network info cache for port 825ca180-3836-41d5-8d91-f55bc2ec9853 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1147.157478] env[62385]: DEBUG oslo_concurrency.lockutils [None req-99ce01a1-a5c3-411a-b4e8-594c02221963 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "28afc06e-f6e4-4044-8bfe-c93b1d9c1388" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.353s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1147.691628] env[62385]: DEBUG nova.network.neutron [req-c59eee15-b5c0-4de8-b332-c7cf96bf15db req-bdf6da5d-b99c-473c-bdad-f7a9ed286cce service nova] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Updated VIF entry in instance network info cache for port 825ca180-3836-41d5-8d91-f55bc2ec9853. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1147.692092] env[62385]: DEBUG nova.network.neutron [req-c59eee15-b5c0-4de8-b332-c7cf96bf15db req-bdf6da5d-b99c-473c-bdad-f7a9ed286cce service nova] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Updating instance_info_cache with network_info: [{"id": "825ca180-3836-41d5-8d91-f55bc2ec9853", "address": "fa:16:3e:e7:99:34", "network": {"id": "06017161-984c-4657-9639-4e6bdb6d4c09", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1072882119-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.208", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d45b5346eab44707b4d81eaaf12b0935", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39cd75b0-9ec7-48ed-b57f-34da0c573a60", "external-id": "nsx-vlan-transportzone-751", "segmentation_id": 751, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap825ca180-38", "ovs_interfaceid": "825ca180-3836-41d5-8d91-f55bc2ec9853", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1147.874208] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "a6a6c7de-656d-4358-977b-f62b471664da" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1147.874563] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "a6a6c7de-656d-4358-977b-f62b471664da" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1148.195796] env[62385]: DEBUG oslo_concurrency.lockutils [req-c59eee15-b5c0-4de8-b332-c7cf96bf15db req-bdf6da5d-b99c-473c-bdad-f7a9ed286cce service nova] Releasing lock "refresh_cache-28afc06e-f6e4-4044-8bfe-c93b1d9c1388" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1148.376815] env[62385]: DEBUG nova.compute.manager [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1148.900762] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1148.901146] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1148.902815] env[62385]: INFO nova.compute.claims [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1149.194799] env[62385]: DEBUG oslo_concurrency.lockutils [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquiring lock "4fe375be-ba60-4601-bd89-610e12742df3" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1149.195197] env[62385]: DEBUG oslo_concurrency.lockutils [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lock "4fe375be-ba60-4601-bd89-610e12742df3" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1149.195458] env[62385]: INFO nova.compute.manager [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Shelving [ 1149.703834] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1149.704095] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cf8acb07-6b2f-428d-819e-357e926b1337 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.711181] env[62385]: DEBUG oslo_vmware.api [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1149.711181] env[62385]: value = "task-1206551" [ 1149.711181] env[62385]: _type = "Task" [ 1149.711181] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.720517] env[62385]: DEBUG oslo_vmware.api [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206551, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.964040] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efc1ec75-80cb-4612-b15e-277cc6c9d879 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.970923] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-342ce164-436c-42df-93c1-5ccced9379b4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.999219] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8afeb80-f923-461b-b845-2b6240304737 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.005605] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec69f9ce-dd18-42c4-967b-faf430f6ca26 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.018148] env[62385]: DEBUG nova.compute.provider_tree [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1150.220456] env[62385]: DEBUG oslo_vmware.api [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206551, 'name': PowerOffVM_Task, 'duration_secs': 0.214662} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.220664] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1150.221422] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6f738ed-29d6-44b2-bd68-299a45d2b379 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.238232] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e2cd1a4-d791-4ab0-bf16-84425f04de28 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.520671] env[62385]: DEBUG nova.scheduler.client.report [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1150.749341] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Creating Snapshot of the VM instance {{(pid=62385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1150.750020] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-537ae3a3-1762-4d02-9581-8efd0e0183ad {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.757996] env[62385]: DEBUG oslo_vmware.api [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1150.757996] env[62385]: value = "task-1206552" [ 1150.757996] env[62385]: _type = "Task" [ 1150.757996] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.765569] env[62385]: DEBUG oslo_vmware.api [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206552, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.025522] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.124s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1151.026054] env[62385]: DEBUG nova.compute.manager [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1151.267700] env[62385]: DEBUG oslo_vmware.api [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206552, 'name': CreateSnapshot_Task, 'duration_secs': 0.431397} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.267950] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Created Snapshot of the VM instance {{(pid=62385) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1151.268655] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bffecdb-157b-485d-b633-e54679220cc8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.531309] env[62385]: DEBUG nova.compute.utils [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1151.532779] env[62385]: DEBUG nova.compute.manager [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1151.532953] env[62385]: DEBUG nova.network.neutron [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1151.571684] env[62385]: DEBUG nova.policy [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4a584f1cc579408eb589777251c6b07b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4831e1f029ed4ffc8ce5bfe7d3f38dd3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 1151.786213] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Creating linked-clone VM from snapshot {{(pid=62385) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1151.786548] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-e90e1460-198c-427d-953e-bd6787f37a5d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.795483] env[62385]: DEBUG oslo_vmware.api [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1151.795483] env[62385]: value = "task-1206553" [ 1151.795483] env[62385]: _type = "Task" [ 1151.795483] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.803370] env[62385]: DEBUG oslo_vmware.api [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206553, 'name': CloneVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.835614] env[62385]: DEBUG nova.network.neutron [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Successfully created port: d861965b-f04d-4e6d-a368-a4d06566f9ae {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1152.036901] env[62385]: DEBUG nova.compute.manager [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1152.306132] env[62385]: DEBUG oslo_vmware.api [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206553, 'name': CloneVM_Task} progress is 94%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.807708] env[62385]: DEBUG oslo_vmware.api [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206553, 'name': CloneVM_Task} progress is 95%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.047135] env[62385]: DEBUG nova.compute.manager [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1153.074839] env[62385]: DEBUG nova.virt.hardware [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1153.075180] env[62385]: DEBUG nova.virt.hardware [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1153.075283] env[62385]: DEBUG nova.virt.hardware [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1153.075440] env[62385]: DEBUG nova.virt.hardware [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1153.075593] env[62385]: DEBUG nova.virt.hardware [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1153.075755] env[62385]: DEBUG nova.virt.hardware [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1153.076112] env[62385]: DEBUG nova.virt.hardware [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1153.076309] env[62385]: DEBUG nova.virt.hardware [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1153.076527] env[62385]: DEBUG nova.virt.hardware [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1153.076743] env[62385]: DEBUG nova.virt.hardware [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1153.077019] env[62385]: DEBUG nova.virt.hardware [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1153.077788] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a19c475-b5a2-4c5f-a4cc-6fe77757b211 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.085805] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a8e0170-d55a-4d23-b9c1-03c914348a90 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.201469] env[62385]: DEBUG nova.compute.manager [req-6f031065-9fdb-4260-9389-d9312e725540 req-4eced17c-c770-420b-8026-0308a91a6c75 service nova] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Received event network-vif-plugged-d861965b-f04d-4e6d-a368-a4d06566f9ae {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1153.201691] env[62385]: DEBUG oslo_concurrency.lockutils [req-6f031065-9fdb-4260-9389-d9312e725540 req-4eced17c-c770-420b-8026-0308a91a6c75 service nova] Acquiring lock "a6a6c7de-656d-4358-977b-f62b471664da-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1153.201920] env[62385]: DEBUG oslo_concurrency.lockutils [req-6f031065-9fdb-4260-9389-d9312e725540 req-4eced17c-c770-420b-8026-0308a91a6c75 service nova] Lock "a6a6c7de-656d-4358-977b-f62b471664da-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1153.202519] env[62385]: DEBUG oslo_concurrency.lockutils [req-6f031065-9fdb-4260-9389-d9312e725540 req-4eced17c-c770-420b-8026-0308a91a6c75 service nova] Lock "a6a6c7de-656d-4358-977b-f62b471664da-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1153.202750] env[62385]: DEBUG nova.compute.manager [req-6f031065-9fdb-4260-9389-d9312e725540 req-4eced17c-c770-420b-8026-0308a91a6c75 service nova] [instance: a6a6c7de-656d-4358-977b-f62b471664da] No waiting events found dispatching network-vif-plugged-d861965b-f04d-4e6d-a368-a4d06566f9ae {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1153.202934] env[62385]: WARNING nova.compute.manager [req-6f031065-9fdb-4260-9389-d9312e725540 req-4eced17c-c770-420b-8026-0308a91a6c75 service nova] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Received unexpected event network-vif-plugged-d861965b-f04d-4e6d-a368-a4d06566f9ae for instance with vm_state building and task_state spawning. [ 1153.309280] env[62385]: DEBUG oslo_vmware.api [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206553, 'name': CloneVM_Task, 'duration_secs': 1.211489} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.309542] env[62385]: INFO nova.virt.vmwareapi.vmops [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Created linked-clone VM from snapshot [ 1153.310353] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51bde51d-6638-4515-b5fc-c7cd9daedab7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.317311] env[62385]: DEBUG nova.virt.vmwareapi.images [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Uploading image a92753a2-4901-40a9-a93b-d846f5d1a602 {{(pid=62385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1153.329631] env[62385]: DEBUG nova.network.neutron [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Successfully updated port: d861965b-f04d-4e6d-a368-a4d06566f9ae {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1153.341547] env[62385]: DEBUG oslo_vmware.rw_handles [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1153.341547] env[62385]: value = "vm-261320" [ 1153.341547] env[62385]: _type = "VirtualMachine" [ 1153.341547] env[62385]: }. {{(pid=62385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1153.341813] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-9a297492-d924-4a04-97ce-95eae1339f3b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.348492] env[62385]: DEBUG oslo_vmware.rw_handles [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lease: (returnval){ [ 1153.348492] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52848201-9920-5d88-a9bb-a6e0121fd348" [ 1153.348492] env[62385]: _type = "HttpNfcLease" [ 1153.348492] env[62385]: } obtained for exporting VM: (result){ [ 1153.348492] env[62385]: value = "vm-261320" [ 1153.348492] env[62385]: _type = "VirtualMachine" [ 1153.348492] env[62385]: }. {{(pid=62385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1153.348793] env[62385]: DEBUG oslo_vmware.api [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the lease: (returnval){ [ 1153.348793] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52848201-9920-5d88-a9bb-a6e0121fd348" [ 1153.348793] env[62385]: _type = "HttpNfcLease" [ 1153.348793] env[62385]: } to be ready. {{(pid=62385) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1153.355351] env[62385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1153.355351] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52848201-9920-5d88-a9bb-a6e0121fd348" [ 1153.355351] env[62385]: _type = "HttpNfcLease" [ 1153.355351] env[62385]: } is initializing. {{(pid=62385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1153.832467] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "refresh_cache-a6a6c7de-656d-4358-977b-f62b471664da" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1153.832634] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquired lock "refresh_cache-a6a6c7de-656d-4358-977b-f62b471664da" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1153.832757] env[62385]: DEBUG nova.network.neutron [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1153.857028] env[62385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1153.857028] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52848201-9920-5d88-a9bb-a6e0121fd348" [ 1153.857028] env[62385]: _type = "HttpNfcLease" [ 1153.857028] env[62385]: } is ready. {{(pid=62385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1153.857419] env[62385]: DEBUG oslo_vmware.rw_handles [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1153.857419] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52848201-9920-5d88-a9bb-a6e0121fd348" [ 1153.857419] env[62385]: _type = "HttpNfcLease" [ 1153.857419] env[62385]: }. {{(pid=62385) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1153.858374] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-555de68c-1eed-43bc-b44d-0afccb9c9481 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.865369] env[62385]: DEBUG oslo_vmware.rw_handles [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5227a9e9-5f53-76ac-22ea-fe7b3a76ca71/disk-0.vmdk from lease info. {{(pid=62385) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1153.865542] env[62385]: DEBUG oslo_vmware.rw_handles [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5227a9e9-5f53-76ac-22ea-fe7b3a76ca71/disk-0.vmdk for reading. {{(pid=62385) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1153.953121] env[62385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-cf3a60d2-6e20-4c0d-9c21-dedfc5a8b4d4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.389064] env[62385]: DEBUG nova.network.neutron [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1154.577201] env[62385]: DEBUG nova.network.neutron [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Updating instance_info_cache with network_info: [{"id": "d861965b-f04d-4e6d-a368-a4d06566f9ae", "address": "fa:16:3e:75:be:11", "network": {"id": "5aef0aec-abdf-4636-9d6b-ba476c7f0410", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1388512761-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4831e1f029ed4ffc8ce5bfe7d3f38dd3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "62f28d75-4e6a-4ae5-b8b3-d0652ea26d08", "external-id": "nsx-vlan-transportzone-801", "segmentation_id": 801, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd861965b-f0", "ovs_interfaceid": "d861965b-f04d-4e6d-a368-a4d06566f9ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1155.080321] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Releasing lock "refresh_cache-a6a6c7de-656d-4358-977b-f62b471664da" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1155.080921] env[62385]: DEBUG nova.compute.manager [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Instance network_info: |[{"id": "d861965b-f04d-4e6d-a368-a4d06566f9ae", "address": "fa:16:3e:75:be:11", "network": {"id": "5aef0aec-abdf-4636-9d6b-ba476c7f0410", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1388512761-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4831e1f029ed4ffc8ce5bfe7d3f38dd3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "62f28d75-4e6a-4ae5-b8b3-d0652ea26d08", "external-id": "nsx-vlan-transportzone-801", "segmentation_id": 801, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd861965b-f0", "ovs_interfaceid": "d861965b-f04d-4e6d-a368-a4d06566f9ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1155.081598] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:75:be:11', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '62f28d75-4e6a-4ae5-b8b3-d0652ea26d08', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd861965b-f04d-4e6d-a368-a4d06566f9ae', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1155.089747] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Creating folder: Project (4831e1f029ed4ffc8ce5bfe7d3f38dd3). Parent ref: group-v261107. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1155.090259] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6a44c78c-a0a3-4194-8040-c1c51434d036 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.102277] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Created folder: Project (4831e1f029ed4ffc8ce5bfe7d3f38dd3) in parent group-v261107. [ 1155.102559] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Creating folder: Instances. Parent ref: group-v261321. {{(pid=62385) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1155.102868] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7414ad7f-9a4d-42ba-bf9b-3cdaac7d0a12 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.113098] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Created folder: Instances in parent group-v261321. [ 1155.113347] env[62385]: DEBUG oslo.service.loopingcall [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1155.113620] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1155.113887] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ed70c6f5-2acf-4cbe-904d-53ed4bb23f90 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.133355] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1155.133355] env[62385]: value = "task-1206557" [ 1155.133355] env[62385]: _type = "Task" [ 1155.133355] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.141627] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206557, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.244318] env[62385]: DEBUG nova.compute.manager [req-b2e19e1c-e47c-4080-ac96-865a350eb036 req-623d9139-bdc6-4829-85cb-39d8efc7699d service nova] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Received event network-changed-d861965b-f04d-4e6d-a368-a4d06566f9ae {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1155.244762] env[62385]: DEBUG nova.compute.manager [req-b2e19e1c-e47c-4080-ac96-865a350eb036 req-623d9139-bdc6-4829-85cb-39d8efc7699d service nova] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Refreshing instance network info cache due to event network-changed-d861965b-f04d-4e6d-a368-a4d06566f9ae. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1155.245086] env[62385]: DEBUG oslo_concurrency.lockutils [req-b2e19e1c-e47c-4080-ac96-865a350eb036 req-623d9139-bdc6-4829-85cb-39d8efc7699d service nova] Acquiring lock "refresh_cache-a6a6c7de-656d-4358-977b-f62b471664da" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1155.245280] env[62385]: DEBUG oslo_concurrency.lockutils [req-b2e19e1c-e47c-4080-ac96-865a350eb036 req-623d9139-bdc6-4829-85cb-39d8efc7699d service nova] Acquired lock "refresh_cache-a6a6c7de-656d-4358-977b-f62b471664da" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1155.245449] env[62385]: DEBUG nova.network.neutron [req-b2e19e1c-e47c-4080-ac96-865a350eb036 req-623d9139-bdc6-4829-85cb-39d8efc7699d service nova] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Refreshing network info cache for port d861965b-f04d-4e6d-a368-a4d06566f9ae {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1155.643513] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206557, 'name': CreateVM_Task, 'duration_secs': 0.333971} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.643822] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1155.644488] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1155.644693] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1155.645186] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1155.645444] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-daef1971-d7de-4b0b-9e7b-c9a0cd4421ac {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.650282] env[62385]: DEBUG oslo_vmware.api [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1155.650282] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]520e362e-0447-5d3c-278a-54a5ae6746cc" [ 1155.650282] env[62385]: _type = "Task" [ 1155.650282] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.659679] env[62385]: DEBUG oslo_vmware.api [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]520e362e-0447-5d3c-278a-54a5ae6746cc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.972217] env[62385]: DEBUG nova.network.neutron [req-b2e19e1c-e47c-4080-ac96-865a350eb036 req-623d9139-bdc6-4829-85cb-39d8efc7699d service nova] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Updated VIF entry in instance network info cache for port d861965b-f04d-4e6d-a368-a4d06566f9ae. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1155.972665] env[62385]: DEBUG nova.network.neutron [req-b2e19e1c-e47c-4080-ac96-865a350eb036 req-623d9139-bdc6-4829-85cb-39d8efc7699d service nova] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Updating instance_info_cache with network_info: [{"id": "d861965b-f04d-4e6d-a368-a4d06566f9ae", "address": "fa:16:3e:75:be:11", "network": {"id": "5aef0aec-abdf-4636-9d6b-ba476c7f0410", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1388512761-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4831e1f029ed4ffc8ce5bfe7d3f38dd3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "62f28d75-4e6a-4ae5-b8b3-d0652ea26d08", "external-id": "nsx-vlan-transportzone-801", "segmentation_id": 801, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd861965b-f0", "ovs_interfaceid": "d861965b-f04d-4e6d-a368-a4d06566f9ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1156.161764] env[62385]: DEBUG oslo_vmware.api [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]520e362e-0447-5d3c-278a-54a5ae6746cc, 'name': SearchDatastore_Task, 'duration_secs': 0.009874} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.162109] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1156.162417] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1156.162677] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1156.162833] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1156.163031] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1156.163323] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-97c21659-eb19-40d4-89ba-70b74d11ecad {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.171255] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1156.171441] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1156.172194] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7223ab3f-5f77-40cb-a1ce-a2cceec226a7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.177465] env[62385]: DEBUG oslo_vmware.api [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1156.177465] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ded3bc-cccb-1c32-2c47-819dd2b83f37" [ 1156.177465] env[62385]: _type = "Task" [ 1156.177465] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.184713] env[62385]: DEBUG oslo_vmware.api [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ded3bc-cccb-1c32-2c47-819dd2b83f37, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.475459] env[62385]: DEBUG oslo_concurrency.lockutils [req-b2e19e1c-e47c-4080-ac96-865a350eb036 req-623d9139-bdc6-4829-85cb-39d8efc7699d service nova] Releasing lock "refresh_cache-a6a6c7de-656d-4358-977b-f62b471664da" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1156.688423] env[62385]: DEBUG oslo_vmware.api [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52ded3bc-cccb-1c32-2c47-819dd2b83f37, 'name': SearchDatastore_Task, 'duration_secs': 0.008942} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.689244] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ede4fbd-d591-46fe-b22d-c4305ce76448 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.694518] env[62385]: DEBUG oslo_vmware.api [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1156.694518] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52f3e7b3-bba1-6ffb-3a87-65a814ce5f89" [ 1156.694518] env[62385]: _type = "Task" [ 1156.694518] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.702353] env[62385]: DEBUG oslo_vmware.api [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52f3e7b3-bba1-6ffb-3a87-65a814ce5f89, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.206044] env[62385]: DEBUG oslo_vmware.api [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52f3e7b3-bba1-6ffb-3a87-65a814ce5f89, 'name': SearchDatastore_Task, 'duration_secs': 0.009058} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.206044] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1157.206244] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] a6a6c7de-656d-4358-977b-f62b471664da/a6a6c7de-656d-4358-977b-f62b471664da.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1157.206428] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d44a8493-dfc5-4c7c-9ca5-be516fb4b4ac {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.212805] env[62385]: DEBUG oslo_vmware.api [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1157.212805] env[62385]: value = "task-1206558" [ 1157.212805] env[62385]: _type = "Task" [ 1157.212805] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.220721] env[62385]: DEBUG oslo_vmware.api [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206558, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.722838] env[62385]: DEBUG oslo_vmware.api [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206558, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.442688} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.723263] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] a6a6c7de-656d-4358-977b-f62b471664da/a6a6c7de-656d-4358-977b-f62b471664da.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1157.723411] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1157.723597] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c82bea2f-232f-4723-8f30-63df161456af {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.730276] env[62385]: DEBUG oslo_vmware.api [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1157.730276] env[62385]: value = "task-1206559" [ 1157.730276] env[62385]: _type = "Task" [ 1157.730276] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.738260] env[62385]: DEBUG oslo_vmware.api [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206559, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.239811] env[62385]: DEBUG oslo_vmware.api [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206559, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066114} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.240159] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1158.240954] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d09e4dc3-8ba9-4cd7-b5fd-df71f483d3b5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.262740] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Reconfiguring VM instance instance-00000073 to attach disk [datastore1] a6a6c7de-656d-4358-977b-f62b471664da/a6a6c7de-656d-4358-977b-f62b471664da.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1158.263082] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f3656b4d-89eb-4834-a3bd-b29bfd3b5649 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.282991] env[62385]: DEBUG oslo_vmware.api [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1158.282991] env[62385]: value = "task-1206560" [ 1158.282991] env[62385]: _type = "Task" [ 1158.282991] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.291906] env[62385]: DEBUG oslo_vmware.api [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206560, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.794342] env[62385]: DEBUG oslo_vmware.api [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206560, 'name': ReconfigVM_Task, 'duration_secs': 0.287569} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.794839] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Reconfigured VM instance instance-00000073 to attach disk [datastore1] a6a6c7de-656d-4358-977b-f62b471664da/a6a6c7de-656d-4358-977b-f62b471664da.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1158.795268] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-565b4a1f-3167-46d3-8241-b01c1c8a41fc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.801589] env[62385]: DEBUG oslo_vmware.api [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1158.801589] env[62385]: value = "task-1206561" [ 1158.801589] env[62385]: _type = "Task" [ 1158.801589] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.809399] env[62385]: DEBUG oslo_vmware.api [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206561, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.311739] env[62385]: DEBUG oslo_vmware.api [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206561, 'name': Rename_Task, 'duration_secs': 0.147322} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.312044] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1159.312349] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c1385989-eef9-456d-9a78-971ac014123a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.319412] env[62385]: DEBUG oslo_vmware.api [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1159.319412] env[62385]: value = "task-1206562" [ 1159.319412] env[62385]: _type = "Task" [ 1159.319412] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.327062] env[62385]: DEBUG oslo_vmware.api [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206562, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.830353] env[62385]: DEBUG oslo_vmware.api [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206562, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.331039] env[62385]: DEBUG oslo_vmware.api [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206562, 'name': PowerOnVM_Task, 'duration_secs': 0.695919} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.331039] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1160.331039] env[62385]: INFO nova.compute.manager [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Took 7.28 seconds to spawn the instance on the hypervisor. [ 1160.331039] env[62385]: DEBUG nova.compute.manager [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1160.331842] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c90d7d6f-abf0-4939-885b-06e37d86ab0a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.851129] env[62385]: INFO nova.compute.manager [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Took 11.97 seconds to build instance. [ 1160.963942] env[62385]: DEBUG oslo_vmware.rw_handles [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5227a9e9-5f53-76ac-22ea-fe7b3a76ca71/disk-0.vmdk. {{(pid=62385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1160.964905] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-579e9dcb-eb91-4047-8297-621f93c5978e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.971635] env[62385]: DEBUG oslo_vmware.rw_handles [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5227a9e9-5f53-76ac-22ea-fe7b3a76ca71/disk-0.vmdk is in state: ready. {{(pid=62385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1160.971815] env[62385]: ERROR oslo_vmware.rw_handles [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5227a9e9-5f53-76ac-22ea-fe7b3a76ca71/disk-0.vmdk due to incomplete transfer. [ 1160.972053] env[62385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-7fd63311-c819-451d-96de-ef6be5580a22 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.979317] env[62385]: DEBUG oslo_vmware.rw_handles [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5227a9e9-5f53-76ac-22ea-fe7b3a76ca71/disk-0.vmdk. {{(pid=62385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1160.979529] env[62385]: DEBUG nova.virt.vmwareapi.images [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Uploaded image a92753a2-4901-40a9-a93b-d846f5d1a602 to the Glance image server {{(pid=62385) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1160.981974] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Destroying the VM {{(pid=62385) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1160.982271] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-c5d78980-7cff-46a7-9e72-ce1e80f2965a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.988617] env[62385]: DEBUG oslo_vmware.api [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1160.988617] env[62385]: value = "task-1206563" [ 1160.988617] env[62385]: _type = "Task" [ 1160.988617] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.996132] env[62385]: DEBUG oslo_vmware.api [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206563, 'name': Destroy_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.355270] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b0ffda49-1866-4a8c-bdbf-33d011afa229 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "a6a6c7de-656d-4358-977b-f62b471664da" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.481s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1161.498675] env[62385]: DEBUG oslo_vmware.api [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206563, 'name': Destroy_Task, 'duration_secs': 0.3057} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.499023] env[62385]: INFO nova.virt.vmwareapi.vm_util [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Destroyed the VM [ 1161.499306] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Deleting Snapshot of the VM instance {{(pid=62385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1161.499579] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-31569202-a886-4352-aa84-f00e66b4fba4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.505788] env[62385]: DEBUG oslo_vmware.api [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1161.505788] env[62385]: value = "task-1206564" [ 1161.505788] env[62385]: _type = "Task" [ 1161.505788] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.513119] env[62385]: DEBUG oslo_vmware.api [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206564, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.016031] env[62385]: DEBUG oslo_vmware.api [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206564, 'name': RemoveSnapshot_Task, 'duration_secs': 0.336637} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.016467] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Deleted Snapshot of the VM instance {{(pid=62385) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1162.016542] env[62385]: DEBUG nova.compute.manager [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1162.017640] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f468d79-5da0-4c83-8af4-0fbba17405fa {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.249117] env[62385]: DEBUG nova.compute.manager [req-b2070dbe-3028-4612-838e-15e209e94cd1 req-e2c6e313-7ff6-4457-b706-9a4e37722271 service nova] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Received event network-changed-d861965b-f04d-4e6d-a368-a4d06566f9ae {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1162.249330] env[62385]: DEBUG nova.compute.manager [req-b2070dbe-3028-4612-838e-15e209e94cd1 req-e2c6e313-7ff6-4457-b706-9a4e37722271 service nova] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Refreshing instance network info cache due to event network-changed-d861965b-f04d-4e6d-a368-a4d06566f9ae. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1162.249559] env[62385]: DEBUG oslo_concurrency.lockutils [req-b2070dbe-3028-4612-838e-15e209e94cd1 req-e2c6e313-7ff6-4457-b706-9a4e37722271 service nova] Acquiring lock "refresh_cache-a6a6c7de-656d-4358-977b-f62b471664da" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1162.249707] env[62385]: DEBUG oslo_concurrency.lockutils [req-b2070dbe-3028-4612-838e-15e209e94cd1 req-e2c6e313-7ff6-4457-b706-9a4e37722271 service nova] Acquired lock "refresh_cache-a6a6c7de-656d-4358-977b-f62b471664da" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1162.249872] env[62385]: DEBUG nova.network.neutron [req-b2070dbe-3028-4612-838e-15e209e94cd1 req-e2c6e313-7ff6-4457-b706-9a4e37722271 service nova] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Refreshing network info cache for port d861965b-f04d-4e6d-a368-a4d06566f9ae {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1162.529481] env[62385]: INFO nova.compute.manager [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Shelve offloading [ 1162.531167] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1162.531448] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bddfc64b-c1a0-40d8-a5d0-d14f13a82197 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.538652] env[62385]: DEBUG oslo_vmware.api [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1162.538652] env[62385]: value = "task-1206565" [ 1162.538652] env[62385]: _type = "Task" [ 1162.538652] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.546372] env[62385]: DEBUG oslo_vmware.api [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206565, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.958413] env[62385]: DEBUG nova.network.neutron [req-b2070dbe-3028-4612-838e-15e209e94cd1 req-e2c6e313-7ff6-4457-b706-9a4e37722271 service nova] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Updated VIF entry in instance network info cache for port d861965b-f04d-4e6d-a368-a4d06566f9ae. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1162.958782] env[62385]: DEBUG nova.network.neutron [req-b2070dbe-3028-4612-838e-15e209e94cd1 req-e2c6e313-7ff6-4457-b706-9a4e37722271 service nova] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Updating instance_info_cache with network_info: [{"id": "d861965b-f04d-4e6d-a368-a4d06566f9ae", "address": "fa:16:3e:75:be:11", "network": {"id": "5aef0aec-abdf-4636-9d6b-ba476c7f0410", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1388512761-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.141", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4831e1f029ed4ffc8ce5bfe7d3f38dd3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "62f28d75-4e6a-4ae5-b8b3-d0652ea26d08", "external-id": "nsx-vlan-transportzone-801", "segmentation_id": 801, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd861965b-f0", "ovs_interfaceid": "d861965b-f04d-4e6d-a368-a4d06566f9ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1163.049258] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] VM already powered off {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1163.049589] env[62385]: DEBUG nova.compute.manager [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1163.050224] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42fc1c96-96df-45da-9535-c14485a7543f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.055962] env[62385]: DEBUG oslo_concurrency.lockutils [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquiring lock "refresh_cache-4fe375be-ba60-4601-bd89-610e12742df3" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1163.056146] env[62385]: DEBUG oslo_concurrency.lockutils [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquired lock "refresh_cache-4fe375be-ba60-4601-bd89-610e12742df3" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1163.056322] env[62385]: DEBUG nova.network.neutron [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1163.461950] env[62385]: DEBUG oslo_concurrency.lockutils [req-b2070dbe-3028-4612-838e-15e209e94cd1 req-e2c6e313-7ff6-4457-b706-9a4e37722271 service nova] Releasing lock "refresh_cache-a6a6c7de-656d-4358-977b-f62b471664da" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1163.749665] env[62385]: DEBUG nova.network.neutron [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Updating instance_info_cache with network_info: [{"id": "1e16fa5f-3cab-4323-8637-7ff6dffce79b", "address": "fa:16:3e:6c:0c:dd", "network": {"id": "9c4d6464-cacc-44f8-b7e5-0699cd312c53", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1889153615-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.148", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09c2c5e938bf46fd9353e47aecc72478", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4d3f69a-b086-4c3b-b976-5a848b63dfc4", "external-id": "nsx-vlan-transportzone-627", "segmentation_id": 627, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e16fa5f-3c", "ovs_interfaceid": "1e16fa5f-3cab-4323-8637-7ff6dffce79b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1164.252664] env[62385]: DEBUG oslo_concurrency.lockutils [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Releasing lock "refresh_cache-4fe375be-ba60-4601-bd89-610e12742df3" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1164.459053] env[62385]: DEBUG nova.compute.manager [req-7e9a5731-a5cf-40c0-80aa-df15b3cb0a05 req-36fc6007-ddf0-44b2-b9b1-ae6f731f5dd6 service nova] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Received event network-vif-unplugged-1e16fa5f-3cab-4323-8637-7ff6dffce79b {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1164.459298] env[62385]: DEBUG oslo_concurrency.lockutils [req-7e9a5731-a5cf-40c0-80aa-df15b3cb0a05 req-36fc6007-ddf0-44b2-b9b1-ae6f731f5dd6 service nova] Acquiring lock "4fe375be-ba60-4601-bd89-610e12742df3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1164.459514] env[62385]: DEBUG oslo_concurrency.lockutils [req-7e9a5731-a5cf-40c0-80aa-df15b3cb0a05 req-36fc6007-ddf0-44b2-b9b1-ae6f731f5dd6 service nova] Lock "4fe375be-ba60-4601-bd89-610e12742df3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1164.459688] env[62385]: DEBUG oslo_concurrency.lockutils [req-7e9a5731-a5cf-40c0-80aa-df15b3cb0a05 req-36fc6007-ddf0-44b2-b9b1-ae6f731f5dd6 service nova] Lock "4fe375be-ba60-4601-bd89-610e12742df3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1164.463333] env[62385]: DEBUG nova.compute.manager [req-7e9a5731-a5cf-40c0-80aa-df15b3cb0a05 req-36fc6007-ddf0-44b2-b9b1-ae6f731f5dd6 service nova] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] No waiting events found dispatching network-vif-unplugged-1e16fa5f-3cab-4323-8637-7ff6dffce79b {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1164.463564] env[62385]: WARNING nova.compute.manager [req-7e9a5731-a5cf-40c0-80aa-df15b3cb0a05 req-36fc6007-ddf0-44b2-b9b1-ae6f731f5dd6 service nova] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Received unexpected event network-vif-unplugged-1e16fa5f-3cab-4323-8637-7ff6dffce79b for instance with vm_state shelved and task_state shelving_offloading. [ 1164.583912] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1164.584878] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdaf0452-3e72-4ba0-a8ad-67c56b02c12b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.592848] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1164.593408] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ebc6db7b-4ec1-4ac2-9b95-d38fa6b95fe8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.659963] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1164.660280] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1164.660492] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Deleting the datastore file [datastore1] 4fe375be-ba60-4601-bd89-610e12742df3 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1164.660767] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-444e38ce-6acc-43ef-bb3b-1e170ebc72fa {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.667224] env[62385]: DEBUG oslo_vmware.api [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1164.667224] env[62385]: value = "task-1206567" [ 1164.667224] env[62385]: _type = "Task" [ 1164.667224] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.674980] env[62385]: DEBUG oslo_vmware.api [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206567, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.177754] env[62385]: DEBUG oslo_vmware.api [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206567, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.127414} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.178050] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1165.178196] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1165.178373] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1165.199342] env[62385]: INFO nova.scheduler.client.report [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Deleted allocations for instance 4fe375be-ba60-4601-bd89-610e12742df3 [ 1165.703916] env[62385]: DEBUG oslo_concurrency.lockutils [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.704228] env[62385]: DEBUG oslo_concurrency.lockutils [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.704295] env[62385]: DEBUG nova.objects.instance [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lazy-loading 'resources' on Instance uuid 4fe375be-ba60-4601-bd89-610e12742df3 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1166.207128] env[62385]: DEBUG nova.objects.instance [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lazy-loading 'numa_topology' on Instance uuid 4fe375be-ba60-4601-bd89-610e12742df3 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1166.486965] env[62385]: DEBUG nova.compute.manager [req-f0285bea-75bf-4c63-8d9f-c0b6b016bd54 req-b7bf63e1-d7bd-4d90-a290-b70d648f28c7 service nova] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Received event network-changed-1e16fa5f-3cab-4323-8637-7ff6dffce79b {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1166.487199] env[62385]: DEBUG nova.compute.manager [req-f0285bea-75bf-4c63-8d9f-c0b6b016bd54 req-b7bf63e1-d7bd-4d90-a290-b70d648f28c7 service nova] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Refreshing instance network info cache due to event network-changed-1e16fa5f-3cab-4323-8637-7ff6dffce79b. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1166.487431] env[62385]: DEBUG oslo_concurrency.lockutils [req-f0285bea-75bf-4c63-8d9f-c0b6b016bd54 req-b7bf63e1-d7bd-4d90-a290-b70d648f28c7 service nova] Acquiring lock "refresh_cache-4fe375be-ba60-4601-bd89-610e12742df3" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1166.487605] env[62385]: DEBUG oslo_concurrency.lockutils [req-f0285bea-75bf-4c63-8d9f-c0b6b016bd54 req-b7bf63e1-d7bd-4d90-a290-b70d648f28c7 service nova] Acquired lock "refresh_cache-4fe375be-ba60-4601-bd89-610e12742df3" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1166.487779] env[62385]: DEBUG nova.network.neutron [req-f0285bea-75bf-4c63-8d9f-c0b6b016bd54 req-b7bf63e1-d7bd-4d90-a290-b70d648f28c7 service nova] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Refreshing network info cache for port 1e16fa5f-3cab-4323-8637-7ff6dffce79b {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1166.711248] env[62385]: DEBUG nova.objects.base [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Object Instance<4fe375be-ba60-4601-bd89-610e12742df3> lazy-loaded attributes: resources,numa_topology {{(pid=62385) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1166.754711] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27932648-d13f-4372-b766-938bd316ec4b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.762834] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-451a60c2-cd30-4ceb-b314-70b987db4344 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.792438] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1588aae4-6e6e-42e3-8a72-1b6961aaf313 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.799541] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bc9e574-d284-4139-8342-08e8dac5808d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.812346] env[62385]: DEBUG nova.compute.provider_tree [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1167.190523] env[62385]: DEBUG nova.network.neutron [req-f0285bea-75bf-4c63-8d9f-c0b6b016bd54 req-b7bf63e1-d7bd-4d90-a290-b70d648f28c7 service nova] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Updated VIF entry in instance network info cache for port 1e16fa5f-3cab-4323-8637-7ff6dffce79b. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1167.191000] env[62385]: DEBUG nova.network.neutron [req-f0285bea-75bf-4c63-8d9f-c0b6b016bd54 req-b7bf63e1-d7bd-4d90-a290-b70d648f28c7 service nova] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Updating instance_info_cache with network_info: [{"id": "1e16fa5f-3cab-4323-8637-7ff6dffce79b", "address": "fa:16:3e:6c:0c:dd", "network": {"id": "9c4d6464-cacc-44f8-b7e5-0699cd312c53", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1889153615-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.148", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09c2c5e938bf46fd9353e47aecc72478", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap1e16fa5f-3c", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1167.314769] env[62385]: DEBUG nova.scheduler.client.report [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1167.694456] env[62385]: DEBUG oslo_concurrency.lockutils [req-f0285bea-75bf-4c63-8d9f-c0b6b016bd54 req-b7bf63e1-d7bd-4d90-a290-b70d648f28c7 service nova] Releasing lock "refresh_cache-4fe375be-ba60-4601-bd89-610e12742df3" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1167.741346] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquiring lock "4fe375be-ba60-4601-bd89-610e12742df3" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1167.818955] env[62385]: DEBUG oslo_concurrency.lockutils [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.115s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1168.327027] env[62385]: DEBUG oslo_concurrency.lockutils [None req-076bdb22-aaf5-42fa-89f2-f74c1c35f15c tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lock "4fe375be-ba60-4601-bd89-610e12742df3" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 19.131s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1168.327027] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lock "4fe375be-ba60-4601-bd89-610e12742df3" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.586s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1168.327027] env[62385]: INFO nova.compute.manager [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Unshelving [ 1169.354429] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1169.354711] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1169.354888] env[62385]: DEBUG nova.objects.instance [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lazy-loading 'pci_requests' on Instance uuid 4fe375be-ba60-4601-bd89-610e12742df3 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1169.858900] env[62385]: DEBUG nova.objects.instance [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lazy-loading 'numa_topology' on Instance uuid 4fe375be-ba60-4601-bd89-610e12742df3 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1170.361390] env[62385]: INFO nova.compute.claims [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1171.416428] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-339e637f-25cc-416e-a10d-b45de38c39cb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.423776] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4ef6929-5ced-4126-b978-e062ef336c18 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.453330] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc347ba9-8516-4da6-b570-bc458ae5c347 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.459661] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a563dca4-b362-4c2a-8dff-2230fa0ee149 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.471866] env[62385]: DEBUG nova.compute.provider_tree [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1171.974744] env[62385]: DEBUG nova.scheduler.client.report [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1172.479715] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.125s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1172.510064] env[62385]: INFO nova.network.neutron [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Updating port 1e16fa5f-3cab-4323-8637-7ff6dffce79b with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1173.892650] env[62385]: DEBUG nova.compute.manager [req-565adbe3-909a-4920-a731-7fda578fc0a8 req-b0e88b52-6add-40cd-9e67-4ec17883e62c service nova] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Received event network-vif-plugged-1e16fa5f-3cab-4323-8637-7ff6dffce79b {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1173.892926] env[62385]: DEBUG oslo_concurrency.lockutils [req-565adbe3-909a-4920-a731-7fda578fc0a8 req-b0e88b52-6add-40cd-9e67-4ec17883e62c service nova] Acquiring lock "4fe375be-ba60-4601-bd89-610e12742df3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1173.893195] env[62385]: DEBUG oslo_concurrency.lockutils [req-565adbe3-909a-4920-a731-7fda578fc0a8 req-b0e88b52-6add-40cd-9e67-4ec17883e62c service nova] Lock "4fe375be-ba60-4601-bd89-610e12742df3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1173.893427] env[62385]: DEBUG oslo_concurrency.lockutils [req-565adbe3-909a-4920-a731-7fda578fc0a8 req-b0e88b52-6add-40cd-9e67-4ec17883e62c service nova] Lock "4fe375be-ba60-4601-bd89-610e12742df3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1173.893608] env[62385]: DEBUG nova.compute.manager [req-565adbe3-909a-4920-a731-7fda578fc0a8 req-b0e88b52-6add-40cd-9e67-4ec17883e62c service nova] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] No waiting events found dispatching network-vif-plugged-1e16fa5f-3cab-4323-8637-7ff6dffce79b {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1173.893783] env[62385]: WARNING nova.compute.manager [req-565adbe3-909a-4920-a731-7fda578fc0a8 req-b0e88b52-6add-40cd-9e67-4ec17883e62c service nova] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Received unexpected event network-vif-plugged-1e16fa5f-3cab-4323-8637-7ff6dffce79b for instance with vm_state shelved_offloaded and task_state spawning. [ 1173.975871] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquiring lock "refresh_cache-4fe375be-ba60-4601-bd89-610e12742df3" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1173.976082] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquired lock "refresh_cache-4fe375be-ba60-4601-bd89-610e12742df3" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1173.976263] env[62385]: DEBUG nova.network.neutron [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1174.703533] env[62385]: DEBUG nova.network.neutron [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Updating instance_info_cache with network_info: [{"id": "1e16fa5f-3cab-4323-8637-7ff6dffce79b", "address": "fa:16:3e:6c:0c:dd", "network": {"id": "9c4d6464-cacc-44f8-b7e5-0699cd312c53", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1889153615-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.148", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09c2c5e938bf46fd9353e47aecc72478", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4d3f69a-b086-4c3b-b976-5a848b63dfc4", "external-id": "nsx-vlan-transportzone-627", "segmentation_id": 627, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e16fa5f-3c", "ovs_interfaceid": "1e16fa5f-3cab-4323-8637-7ff6dffce79b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1175.206605] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Releasing lock "refresh_cache-4fe375be-ba60-4601-bd89-610e12742df3" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1175.232653] env[62385]: DEBUG nova.virt.hardware [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='15ac3869f532d0ca4fac13b86f01ac69',container_format='bare',created_at=2024-10-09T02:28:20Z,direct_url=,disk_format='vmdk',id=a92753a2-4901-40a9-a93b-d846f5d1a602,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1219976290-shelved',owner='09c2c5e938bf46fd9353e47aecc72478',properties=ImageMetaProps,protected=,size=31669248,status='active',tags=,updated_at=2024-10-09T02:28:32Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1175.232901] env[62385]: DEBUG nova.virt.hardware [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1175.233079] env[62385]: DEBUG nova.virt.hardware [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1175.233277] env[62385]: DEBUG nova.virt.hardware [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1175.233463] env[62385]: DEBUG nova.virt.hardware [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1175.233624] env[62385]: DEBUG nova.virt.hardware [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1175.233836] env[62385]: DEBUG nova.virt.hardware [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1175.234010] env[62385]: DEBUG nova.virt.hardware [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1175.234204] env[62385]: DEBUG nova.virt.hardware [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1175.234373] env[62385]: DEBUG nova.virt.hardware [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1175.234549] env[62385]: DEBUG nova.virt.hardware [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1175.235427] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dce7741-0d08-4d87-b9f3-353d72b100f0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.243078] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68c340ea-ceaf-4b76-a37a-b2259b7954d3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.255648] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6c:0c:dd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c4d3f69a-b086-4c3b-b976-5a848b63dfc4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1e16fa5f-3cab-4323-8637-7ff6dffce79b', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1175.262896] env[62385]: DEBUG oslo.service.loopingcall [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1175.263193] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1175.263345] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4ea1b26f-bae9-4044-a915-bfcc2c056de8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.280982] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1175.280982] env[62385]: value = "task-1206568" [ 1175.280982] env[62385]: _type = "Task" [ 1175.280982] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.288571] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206568, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.791654] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206568, 'name': CreateVM_Task, 'duration_secs': 0.302483} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.791802] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1175.792521] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a92753a2-4901-40a9-a93b-d846f5d1a602" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1175.792702] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a92753a2-4901-40a9-a93b-d846f5d1a602" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1175.793154] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a92753a2-4901-40a9-a93b-d846f5d1a602" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1175.793416] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2369813b-309e-47e0-9e7f-55e154fb7052 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.797589] env[62385]: DEBUG oslo_vmware.api [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1175.797589] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52f5a575-44fa-0ee1-a846-0e3d37a5c10e" [ 1175.797589] env[62385]: _type = "Task" [ 1175.797589] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.806145] env[62385]: DEBUG oslo_vmware.api [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52f5a575-44fa-0ee1-a846-0e3d37a5c10e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.915251] env[62385]: DEBUG nova.compute.manager [req-55dae6cb-1c4b-4c4c-8edf-9b35ab53f2bb req-a67bec13-9f71-46c5-8e3e-9912b1295f87 service nova] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Received event network-changed-1e16fa5f-3cab-4323-8637-7ff6dffce79b {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1175.915447] env[62385]: DEBUG nova.compute.manager [req-55dae6cb-1c4b-4c4c-8edf-9b35ab53f2bb req-a67bec13-9f71-46c5-8e3e-9912b1295f87 service nova] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Refreshing instance network info cache due to event network-changed-1e16fa5f-3cab-4323-8637-7ff6dffce79b. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1175.915671] env[62385]: DEBUG oslo_concurrency.lockutils [req-55dae6cb-1c4b-4c4c-8edf-9b35ab53f2bb req-a67bec13-9f71-46c5-8e3e-9912b1295f87 service nova] Acquiring lock "refresh_cache-4fe375be-ba60-4601-bd89-610e12742df3" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1175.915850] env[62385]: DEBUG oslo_concurrency.lockutils [req-55dae6cb-1c4b-4c4c-8edf-9b35ab53f2bb req-a67bec13-9f71-46c5-8e3e-9912b1295f87 service nova] Acquired lock "refresh_cache-4fe375be-ba60-4601-bd89-610e12742df3" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1175.915988] env[62385]: DEBUG nova.network.neutron [req-55dae6cb-1c4b-4c4c-8edf-9b35ab53f2bb req-a67bec13-9f71-46c5-8e3e-9912b1295f87 service nova] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Refreshing network info cache for port 1e16fa5f-3cab-4323-8637-7ff6dffce79b {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1176.307581] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a92753a2-4901-40a9-a93b-d846f5d1a602" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1176.307895] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Processing image a92753a2-4901-40a9-a93b-d846f5d1a602 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1176.308034] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a92753a2-4901-40a9-a93b-d846f5d1a602/a92753a2-4901-40a9-a93b-d846f5d1a602.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1176.308199] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a92753a2-4901-40a9-a93b-d846f5d1a602/a92753a2-4901-40a9-a93b-d846f5d1a602.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1176.308388] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1176.308680] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8f254ced-77fe-46f7-8632-0889c5a38011 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.326107] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1176.326288] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1176.326980] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-978a180a-9d25-4e7c-bdec-45c3ff94a1df {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.331798] env[62385]: DEBUG oslo_vmware.api [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1176.331798] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52853e90-fd7c-6fb6-9b4c-91b2bca0b104" [ 1176.331798] env[62385]: _type = "Task" [ 1176.331798] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.338822] env[62385]: DEBUG oslo_vmware.api [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52853e90-fd7c-6fb6-9b4c-91b2bca0b104, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.612771] env[62385]: DEBUG nova.network.neutron [req-55dae6cb-1c4b-4c4c-8edf-9b35ab53f2bb req-a67bec13-9f71-46c5-8e3e-9912b1295f87 service nova] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Updated VIF entry in instance network info cache for port 1e16fa5f-3cab-4323-8637-7ff6dffce79b. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1176.613142] env[62385]: DEBUG nova.network.neutron [req-55dae6cb-1c4b-4c4c-8edf-9b35ab53f2bb req-a67bec13-9f71-46c5-8e3e-9912b1295f87 service nova] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Updating instance_info_cache with network_info: [{"id": "1e16fa5f-3cab-4323-8637-7ff6dffce79b", "address": "fa:16:3e:6c:0c:dd", "network": {"id": "9c4d6464-cacc-44f8-b7e5-0699cd312c53", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1889153615-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.148", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "09c2c5e938bf46fd9353e47aecc72478", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4d3f69a-b086-4c3b-b976-5a848b63dfc4", "external-id": "nsx-vlan-transportzone-627", "segmentation_id": 627, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e16fa5f-3c", "ovs_interfaceid": "1e16fa5f-3cab-4323-8637-7ff6dffce79b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1176.841586] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Preparing fetch location {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1176.841858] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Fetch image to [datastore2] OSTACK_IMG_bb3389d5-8c92-46fd-b00c-f5cada900404/OSTACK_IMG_bb3389d5-8c92-46fd-b00c-f5cada900404.vmdk {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1176.842066] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Downloading stream optimized image a92753a2-4901-40a9-a93b-d846f5d1a602 to [datastore2] OSTACK_IMG_bb3389d5-8c92-46fd-b00c-f5cada900404/OSTACK_IMG_bb3389d5-8c92-46fd-b00c-f5cada900404.vmdk on the data store datastore2 as vApp {{(pid=62385) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1176.842247] env[62385]: DEBUG nova.virt.vmwareapi.images [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Downloading image file data a92753a2-4901-40a9-a93b-d846f5d1a602 to the ESX as VM named 'OSTACK_IMG_bb3389d5-8c92-46fd-b00c-f5cada900404' {{(pid=62385) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1176.907390] env[62385]: DEBUG oslo_vmware.rw_handles [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1176.907390] env[62385]: value = "resgroup-9" [ 1176.907390] env[62385]: _type = "ResourcePool" [ 1176.907390] env[62385]: }. {{(pid=62385) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1176.907696] env[62385]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-cb631d2c-2365-4c7c-8657-c7c706ef4f8a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.928226] env[62385]: DEBUG oslo_vmware.rw_handles [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lease: (returnval){ [ 1176.928226] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e565ae-aab0-49ea-4e81-858e804c6196" [ 1176.928226] env[62385]: _type = "HttpNfcLease" [ 1176.928226] env[62385]: } obtained for vApp import into resource pool (val){ [ 1176.928226] env[62385]: value = "resgroup-9" [ 1176.928226] env[62385]: _type = "ResourcePool" [ 1176.928226] env[62385]: }. {{(pid=62385) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1176.928551] env[62385]: DEBUG oslo_vmware.api [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the lease: (returnval){ [ 1176.928551] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e565ae-aab0-49ea-4e81-858e804c6196" [ 1176.928551] env[62385]: _type = "HttpNfcLease" [ 1176.928551] env[62385]: } to be ready. {{(pid=62385) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1176.934524] env[62385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1176.934524] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e565ae-aab0-49ea-4e81-858e804c6196" [ 1176.934524] env[62385]: _type = "HttpNfcLease" [ 1176.934524] env[62385]: } is initializing. {{(pid=62385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1177.116510] env[62385]: DEBUG oslo_concurrency.lockutils [req-55dae6cb-1c4b-4c4c-8edf-9b35ab53f2bb req-a67bec13-9f71-46c5-8e3e-9912b1295f87 service nova] Releasing lock "refresh_cache-4fe375be-ba60-4601-bd89-610e12742df3" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1177.436645] env[62385]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1177.436645] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e565ae-aab0-49ea-4e81-858e804c6196" [ 1177.436645] env[62385]: _type = "HttpNfcLease" [ 1177.436645] env[62385]: } is ready. {{(pid=62385) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1177.437123] env[62385]: DEBUG oslo_vmware.rw_handles [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1177.437123] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52e565ae-aab0-49ea-4e81-858e804c6196" [ 1177.437123] env[62385]: _type = "HttpNfcLease" [ 1177.437123] env[62385]: }. {{(pid=62385) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1177.437619] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ef81651-5361-42cb-9b46-c4cd23018112 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.444573] env[62385]: DEBUG oslo_vmware.rw_handles [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5223ae53-9d15-79c7-3b60-de86503a048a/disk-0.vmdk from lease info. {{(pid=62385) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1177.444749] env[62385]: DEBUG oslo_vmware.rw_handles [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Creating HTTP connection to write to file with size = 31669248 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5223ae53-9d15-79c7-3b60-de86503a048a/disk-0.vmdk. {{(pid=62385) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1177.507272] env[62385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-0fe4a074-f6e8-4e03-9374-4d503fed0aec {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.744137] env[62385]: DEBUG oslo_vmware.rw_handles [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Completed reading data from the image iterator. {{(pid=62385) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1178.744586] env[62385]: DEBUG oslo_vmware.rw_handles [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5223ae53-9d15-79c7-3b60-de86503a048a/disk-0.vmdk. {{(pid=62385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1178.745465] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58e65379-eb26-4ddc-becb-5c08e2ecaed5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.752592] env[62385]: DEBUG oslo_vmware.rw_handles [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5223ae53-9d15-79c7-3b60-de86503a048a/disk-0.vmdk is in state: ready. {{(pid=62385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1178.752800] env[62385]: DEBUG oslo_vmware.rw_handles [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5223ae53-9d15-79c7-3b60-de86503a048a/disk-0.vmdk. {{(pid=62385) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1178.753049] env[62385]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-6c9383b6-100b-4266-b78a-f8d9181cea93 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.978445] env[62385]: DEBUG oslo_vmware.rw_handles [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5223ae53-9d15-79c7-3b60-de86503a048a/disk-0.vmdk. {{(pid=62385) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1178.978718] env[62385]: INFO nova.virt.vmwareapi.images [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Downloaded image file data a92753a2-4901-40a9-a93b-d846f5d1a602 [ 1178.979519] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ffdf69d-4c41-4c5f-b729-064f6ac8fe1f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.994670] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2135adb4-b311-4376-aa4c-5db571014d9d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.062419] env[62385]: INFO nova.virt.vmwareapi.images [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] The imported VM was unregistered [ 1179.064831] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Caching image {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1179.065084] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Creating directory with path [datastore2] devstack-image-cache_base/a92753a2-4901-40a9-a93b-d846f5d1a602 {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1179.065358] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-62065788-a983-43fe-bfd6-4e457e0ac5ee {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.075548] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Created directory with path [datastore2] devstack-image-cache_base/a92753a2-4901-40a9-a93b-d846f5d1a602 {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1179.075733] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_bb3389d5-8c92-46fd-b00c-f5cada900404/OSTACK_IMG_bb3389d5-8c92-46fd-b00c-f5cada900404.vmdk to [datastore2] devstack-image-cache_base/a92753a2-4901-40a9-a93b-d846f5d1a602/a92753a2-4901-40a9-a93b-d846f5d1a602.vmdk. {{(pid=62385) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1179.075972] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-88e4417c-fcd5-4cfd-9dbf-92ad197ae5ed {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.082123] env[62385]: DEBUG oslo_vmware.api [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1179.082123] env[62385]: value = "task-1206571" [ 1179.082123] env[62385]: _type = "Task" [ 1179.082123] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.090483] env[62385]: DEBUG oslo_vmware.api [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206571, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.593884] env[62385]: DEBUG oslo_vmware.api [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206571, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.094953] env[62385]: DEBUG oslo_vmware.api [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206571, 'name': MoveVirtualDisk_Task} progress is 49%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.596871] env[62385]: DEBUG oslo_vmware.api [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206571, 'name': MoveVirtualDisk_Task} progress is 74%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.097873] env[62385]: DEBUG oslo_vmware.api [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206571, 'name': MoveVirtualDisk_Task} progress is 97%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.596420] env[62385]: DEBUG oslo_vmware.api [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206571, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.130404} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.596689] env[62385]: INFO nova.virt.vmwareapi.ds_util [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_bb3389d5-8c92-46fd-b00c-f5cada900404/OSTACK_IMG_bb3389d5-8c92-46fd-b00c-f5cada900404.vmdk to [datastore2] devstack-image-cache_base/a92753a2-4901-40a9-a93b-d846f5d1a602/a92753a2-4901-40a9-a93b-d846f5d1a602.vmdk. [ 1181.596869] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Cleaning up location [datastore2] OSTACK_IMG_bb3389d5-8c92-46fd-b00c-f5cada900404 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1181.597048] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_bb3389d5-8c92-46fd-b00c-f5cada900404 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1181.597304] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-20f594dd-e9d0-4a77-8a7d-7c950540be3f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.603561] env[62385]: DEBUG oslo_vmware.api [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1181.603561] env[62385]: value = "task-1206572" [ 1181.603561] env[62385]: _type = "Task" [ 1181.603561] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.611403] env[62385]: DEBUG oslo_vmware.api [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206572, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.114049] env[62385]: DEBUG oslo_vmware.api [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206572, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.057864} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.114049] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1182.114049] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a92753a2-4901-40a9-a93b-d846f5d1a602/a92753a2-4901-40a9-a93b-d846f5d1a602.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1182.114049] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a92753a2-4901-40a9-a93b-d846f5d1a602/a92753a2-4901-40a9-a93b-d846f5d1a602.vmdk to [datastore2] 4fe375be-ba60-4601-bd89-610e12742df3/4fe375be-ba60-4601-bd89-610e12742df3.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1182.114563] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0bc50b5e-8898-4565-87c3-66002e588d82 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.120530] env[62385]: DEBUG oslo_vmware.api [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1182.120530] env[62385]: value = "task-1206573" [ 1182.120530] env[62385]: _type = "Task" [ 1182.120530] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.127498] env[62385]: DEBUG oslo_vmware.api [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206573, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.632929] env[62385]: DEBUG oslo_vmware.api [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206573, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.133332] env[62385]: DEBUG oslo_vmware.api [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206573, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.634434] env[62385]: DEBUG oslo_vmware.api [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206573, 'name': CopyVirtualDisk_Task} progress is 71%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.135356] env[62385]: DEBUG oslo_vmware.api [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206573, 'name': CopyVirtualDisk_Task} progress is 91%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.388625] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1f881097-97eb-4c52-8f02-85c471b3134e tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "28afc06e-f6e4-4044-8bfe-c93b1d9c1388" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1184.388861] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1f881097-97eb-4c52-8f02-85c471b3134e tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "28afc06e-f6e4-4044-8bfe-c93b1d9c1388" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1184.389072] env[62385]: DEBUG nova.compute.manager [None req-1f881097-97eb-4c52-8f02-85c471b3134e tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1184.389984] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b2dd1cc-53fd-48de-b44b-7ea2dd019d85 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.396664] env[62385]: DEBUG nova.compute.manager [None req-1f881097-97eb-4c52-8f02-85c471b3134e tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62385) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1184.397250] env[62385]: DEBUG nova.objects.instance [None req-1f881097-97eb-4c52-8f02-85c471b3134e tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lazy-loading 'flavor' on Instance uuid 28afc06e-f6e4-4044-8bfe-c93b1d9c1388 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1184.633479] env[62385]: DEBUG oslo_vmware.api [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206573, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.153728} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.633771] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a92753a2-4901-40a9-a93b-d846f5d1a602/a92753a2-4901-40a9-a93b-d846f5d1a602.vmdk to [datastore2] 4fe375be-ba60-4601-bd89-610e12742df3/4fe375be-ba60-4601-bd89-610e12742df3.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1184.634585] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fd96b93-717e-4632-b983-c9cc3b51ecc0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.657258] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Reconfiguring VM instance instance-00000071 to attach disk [datastore2] 4fe375be-ba60-4601-bd89-610e12742df3/4fe375be-ba60-4601-bd89-610e12742df3.vmdk or device None with type streamOptimized {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1184.657477] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-91ca70fe-3dcc-4a57-91cd-f01bb7cd25b7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.675468] env[62385]: DEBUG oslo_vmware.api [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1184.675468] env[62385]: value = "task-1206574" [ 1184.675468] env[62385]: _type = "Task" [ 1184.675468] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.682802] env[62385]: DEBUG oslo_vmware.api [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206574, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.902704] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f881097-97eb-4c52-8f02-85c471b3134e tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1184.902974] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3ad968a7-1107-4226-8f9f-cf8955ebefd5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.910459] env[62385]: DEBUG oslo_vmware.api [None req-1f881097-97eb-4c52-8f02-85c471b3134e tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1184.910459] env[62385]: value = "task-1206575" [ 1184.910459] env[62385]: _type = "Task" [ 1184.910459] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.918867] env[62385]: DEBUG oslo_vmware.api [None req-1f881097-97eb-4c52-8f02-85c471b3134e tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206575, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.185262] env[62385]: DEBUG oslo_vmware.api [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206574, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.421126] env[62385]: DEBUG oslo_vmware.api [None req-1f881097-97eb-4c52-8f02-85c471b3134e tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206575, 'name': PowerOffVM_Task, 'duration_secs': 0.214185} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.421410] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f881097-97eb-4c52-8f02-85c471b3134e tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1185.421591] env[62385]: DEBUG nova.compute.manager [None req-1f881097-97eb-4c52-8f02-85c471b3134e tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1185.422353] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dac77a6-b922-4738-816d-9e8fc97d2dcd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.687108] env[62385]: DEBUG oslo_vmware.api [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206574, 'name': ReconfigVM_Task, 'duration_secs': 0.525842} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.687434] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Reconfigured VM instance instance-00000071 to attach disk [datastore2] 4fe375be-ba60-4601-bd89-610e12742df3/4fe375be-ba60-4601-bd89-610e12742df3.vmdk or device None with type streamOptimized {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1185.688075] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-70e1729e-4dd4-4428-808e-3307f77d1c11 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.693963] env[62385]: DEBUG oslo_vmware.api [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1185.693963] env[62385]: value = "task-1206576" [ 1185.693963] env[62385]: _type = "Task" [ 1185.693963] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.701191] env[62385]: DEBUG oslo_vmware.api [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206576, 'name': Rename_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.760404] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1185.760670] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1185.760788] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Starting heal instance info cache {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1185.933553] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1f881097-97eb-4c52-8f02-85c471b3134e tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "28afc06e-f6e4-4044-8bfe-c93b1d9c1388" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.544s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1186.204494] env[62385]: DEBUG oslo_vmware.api [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206576, 'name': Rename_Task, 'duration_secs': 0.158555} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.204815] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1186.205026] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e0dea7a8-9692-42d1-af75-33934c97e351 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.212337] env[62385]: DEBUG oslo_vmware.api [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1186.212337] env[62385]: value = "task-1206577" [ 1186.212337] env[62385]: _type = "Task" [ 1186.212337] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.219670] env[62385]: DEBUG oslo_vmware.api [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206577, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.722885] env[62385]: DEBUG oslo_vmware.api [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206577, 'name': PowerOnVM_Task, 'duration_secs': 0.428466} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.723198] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1186.765919] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Didn't find any instances for network info cache update. {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1186.766246] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1186.766541] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1186.766741] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1186.766930] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1186.767124] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1186.767288] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62385) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1186.786340] env[62385]: DEBUG nova.objects.instance [None req-ee55f03c-6fda-4b0d-ae42-abeba89a44fa tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lazy-loading 'flavor' on Instance uuid 28afc06e-f6e4-4044-8bfe-c93b1d9c1388 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1186.821140] env[62385]: DEBUG nova.compute.manager [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1186.822089] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4daae39b-f5ee-4c8c-a77d-32f245d105fa {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.030496] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1187.291399] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ee55f03c-6fda-4b0d-ae42-abeba89a44fa tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "refresh_cache-28afc06e-f6e4-4044-8bfe-c93b1d9c1388" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1187.291760] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ee55f03c-6fda-4b0d-ae42-abeba89a44fa tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquired lock "refresh_cache-28afc06e-f6e4-4044-8bfe-c93b1d9c1388" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1187.291760] env[62385]: DEBUG nova.network.neutron [None req-ee55f03c-6fda-4b0d-ae42-abeba89a44fa tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1187.291901] env[62385]: DEBUG nova.objects.instance [None req-ee55f03c-6fda-4b0d-ae42-abeba89a44fa tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lazy-loading 'info_cache' on Instance uuid 28afc06e-f6e4-4044-8bfe-c93b1d9c1388 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1187.338968] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4aca8ff4-bdd8-4fd9-a83a-bebd136a4363 tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lock "4fe375be-ba60-4601-bd89-610e12742df3" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 19.012s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1187.795710] env[62385]: DEBUG nova.objects.base [None req-ee55f03c-6fda-4b0d-ae42-abeba89a44fa tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Object Instance<28afc06e-f6e4-4044-8bfe-c93b1d9c1388> lazy-loaded attributes: flavor,info_cache {{(pid=62385) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1188.029599] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager.update_available_resource {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1188.495396] env[62385]: DEBUG nova.network.neutron [None req-ee55f03c-6fda-4b0d-ae42-abeba89a44fa tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Updating instance_info_cache with network_info: [{"id": "825ca180-3836-41d5-8d91-f55bc2ec9853", "address": "fa:16:3e:e7:99:34", "network": {"id": "06017161-984c-4657-9639-4e6bdb6d4c09", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1072882119-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.208", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d45b5346eab44707b4d81eaaf12b0935", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39cd75b0-9ec7-48ed-b57f-34da0c573a60", "external-id": "nsx-vlan-transportzone-751", "segmentation_id": 751, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap825ca180-38", "ovs_interfaceid": "825ca180-3836-41d5-8d91-f55bc2ec9853", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1188.532427] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1188.532688] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1188.532858] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1188.533013] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62385) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1188.533970] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff65cdb7-ed9e-4009-a337-f2ed77a0c11e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.542628] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-628a887d-fefc-4d8a-a203-4b179f4e38d5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.555822] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dd88b53-89aa-4b9b-a8ab-8d701dbe98fb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.561656] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ee4c26c-1bdd-43ef-8872-e01455511d5e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.591571] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181171MB free_disk=155GB free_vcpus=48 pci_devices=None {{(pid=62385) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1188.591736] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1188.591887] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1188.998537] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ee55f03c-6fda-4b0d-ae42-abeba89a44fa tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Releasing lock "refresh_cache-28afc06e-f6e4-4044-8bfe-c93b1d9c1388" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1189.501644] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee55f03c-6fda-4b0d-ae42-abeba89a44fa tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1189.502108] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-97b0e6c7-0cfe-48a1-b78e-90c1fdfb33af {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.509233] env[62385]: DEBUG oslo_vmware.api [None req-ee55f03c-6fda-4b0d-ae42-abeba89a44fa tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1189.509233] env[62385]: value = "task-1206578" [ 1189.509233] env[62385]: _type = "Task" [ 1189.509233] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.516618] env[62385]: DEBUG oslo_vmware.api [None req-ee55f03c-6fda-4b0d-ae42-abeba89a44fa tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206578, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.617256] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 28afc06e-f6e4-4044-8bfe-c93b1d9c1388 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1189.617474] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance a6a6c7de-656d-4358-977b-f62b471664da actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1189.617556] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 4fe375be-ba60-4601-bd89-610e12742df3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1189.617709] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=62385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1189.617849] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=62385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1189.676532] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c000b54-af54-44d4-af9a-d69b59813d43 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.684342] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8af3f2d-2300-4042-b583-e0eacfac03cb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.713465] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bde9fb65-9346-4c78-99ce-a279d48ddde3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.720954] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2a2b1e1-2985-4dc1-adc4-dc44ea42442b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.735814] env[62385]: DEBUG nova.compute.provider_tree [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1190.019335] env[62385]: DEBUG oslo_vmware.api [None req-ee55f03c-6fda-4b0d-ae42-abeba89a44fa tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206578, 'name': PowerOnVM_Task, 'duration_secs': 0.349914} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.019590] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee55f03c-6fda-4b0d-ae42-abeba89a44fa tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1190.019781] env[62385]: DEBUG nova.compute.manager [None req-ee55f03c-6fda-4b0d-ae42-abeba89a44fa tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1190.020561] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7c8cc5a-b1e7-4855-a5e1-44cc51e3a769 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.238509] env[62385]: DEBUG nova.scheduler.client.report [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1190.743870] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62385) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1190.744090] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.152s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1191.793577] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c7f47be-e686-41db-bd89-adc71f366750 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.800350] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-354a5c40-7440-433c-87db-cd912876e5f9 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Suspending the VM {{(pid=62385) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 1191.800653] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-05d7e2e9-2b94-43ec-b049-fb8d50a327ef {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.807317] env[62385]: DEBUG oslo_vmware.api [None req-354a5c40-7440-433c-87db-cd912876e5f9 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1191.807317] env[62385]: value = "task-1206579" [ 1191.807317] env[62385]: _type = "Task" [ 1191.807317] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.816271] env[62385]: DEBUG oslo_vmware.api [None req-354a5c40-7440-433c-87db-cd912876e5f9 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206579, 'name': SuspendVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.317767] env[62385]: DEBUG oslo_vmware.api [None req-354a5c40-7440-433c-87db-cd912876e5f9 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206579, 'name': SuspendVM_Task} progress is 70%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.817626] env[62385]: DEBUG oslo_vmware.api [None req-354a5c40-7440-433c-87db-cd912876e5f9 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206579, 'name': SuspendVM_Task, 'duration_secs': 0.65382} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.818019] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-354a5c40-7440-433c-87db-cd912876e5f9 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Suspended the VM {{(pid=62385) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 1192.818100] env[62385]: DEBUG nova.compute.manager [None req-354a5c40-7440-433c-87db-cd912876e5f9 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1192.818850] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03a3cc36-0218-4f5b-bc0c-840ad3d2ba91 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.138650] env[62385]: INFO nova.compute.manager [None req-4a9a0441-b903-46b8-a1d0-2b123716372c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Resuming [ 1194.139401] env[62385]: DEBUG nova.objects.instance [None req-4a9a0441-b903-46b8-a1d0-2b123716372c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lazy-loading 'flavor' on Instance uuid 28afc06e-f6e4-4044-8bfe-c93b1d9c1388 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1194.739906] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1195.146858] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4a9a0441-b903-46b8-a1d0-2b123716372c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "refresh_cache-28afc06e-f6e4-4044-8bfe-c93b1d9c1388" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1195.147257] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4a9a0441-b903-46b8-a1d0-2b123716372c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquired lock "refresh_cache-28afc06e-f6e4-4044-8bfe-c93b1d9c1388" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1195.147257] env[62385]: DEBUG nova.network.neutron [None req-4a9a0441-b903-46b8-a1d0-2b123716372c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1195.844895] env[62385]: DEBUG nova.network.neutron [None req-4a9a0441-b903-46b8-a1d0-2b123716372c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Updating instance_info_cache with network_info: [{"id": "825ca180-3836-41d5-8d91-f55bc2ec9853", "address": "fa:16:3e:e7:99:34", "network": {"id": "06017161-984c-4657-9639-4e6bdb6d4c09", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1072882119-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.208", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d45b5346eab44707b4d81eaaf12b0935", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39cd75b0-9ec7-48ed-b57f-34da0c573a60", "external-id": "nsx-vlan-transportzone-751", "segmentation_id": 751, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap825ca180-38", "ovs_interfaceid": "825ca180-3836-41d5-8d91-f55bc2ec9853", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1196.347430] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4a9a0441-b903-46b8-a1d0-2b123716372c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Releasing lock "refresh_cache-28afc06e-f6e4-4044-8bfe-c93b1d9c1388" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1196.348547] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c71bd62d-6a95-4d8d-b912-e87367c0355d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.355335] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4a9a0441-b903-46b8-a1d0-2b123716372c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Resuming the VM {{(pid=62385) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1184}} [ 1196.355565] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-82823ad3-82a7-4ac0-809e-ff266305557f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.362011] env[62385]: DEBUG oslo_vmware.api [None req-4a9a0441-b903-46b8-a1d0-2b123716372c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1196.362011] env[62385]: value = "task-1206580" [ 1196.362011] env[62385]: _type = "Task" [ 1196.362011] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.369156] env[62385]: DEBUG oslo_vmware.api [None req-4a9a0441-b903-46b8-a1d0-2b123716372c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206580, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.875128] env[62385]: DEBUG oslo_vmware.api [None req-4a9a0441-b903-46b8-a1d0-2b123716372c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206580, 'name': PowerOnVM_Task, 'duration_secs': 0.499841} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.875503] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4a9a0441-b903-46b8-a1d0-2b123716372c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Resumed the VM {{(pid=62385) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1189}} [ 1196.875635] env[62385]: DEBUG nova.compute.manager [None req-4a9a0441-b903-46b8-a1d0-2b123716372c tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1196.876439] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3989f231-4781-487f-bc24-b06074790a73 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.735561] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1b5fafe9-28f2-419a-831a-3b8c5e3e0623 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "28afc06e-f6e4-4044-8bfe-c93b1d9c1388" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1197.735995] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1b5fafe9-28f2-419a-831a-3b8c5e3e0623 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "28afc06e-f6e4-4044-8bfe-c93b1d9c1388" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1197.735995] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1b5fafe9-28f2-419a-831a-3b8c5e3e0623 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "28afc06e-f6e4-4044-8bfe-c93b1d9c1388-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1197.737607] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1b5fafe9-28f2-419a-831a-3b8c5e3e0623 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "28afc06e-f6e4-4044-8bfe-c93b1d9c1388-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.002s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1197.737799] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1b5fafe9-28f2-419a-831a-3b8c5e3e0623 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "28afc06e-f6e4-4044-8bfe-c93b1d9c1388-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1197.740008] env[62385]: INFO nova.compute.manager [None req-1b5fafe9-28f2-419a-831a-3b8c5e3e0623 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Terminating instance [ 1197.741890] env[62385]: DEBUG nova.compute.manager [None req-1b5fafe9-28f2-419a-831a-3b8c5e3e0623 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1197.742111] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-1b5fafe9-28f2-419a-831a-3b8c5e3e0623 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1197.743047] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67434c4c-4b96-4d73-ae17-0ed2b469a1c1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.750527] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b5fafe9-28f2-419a-831a-3b8c5e3e0623 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1197.750755] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e7a1ecdf-ca4f-46c3-88c9-9af501994940 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.757079] env[62385]: DEBUG oslo_vmware.api [None req-1b5fafe9-28f2-419a-831a-3b8c5e3e0623 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1197.757079] env[62385]: value = "task-1206581" [ 1197.757079] env[62385]: _type = "Task" [ 1197.757079] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1197.764278] env[62385]: DEBUG oslo_vmware.api [None req-1b5fafe9-28f2-419a-831a-3b8c5e3e0623 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206581, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.267574] env[62385]: DEBUG oslo_vmware.api [None req-1b5fafe9-28f2-419a-831a-3b8c5e3e0623 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206581, 'name': PowerOffVM_Task, 'duration_secs': 0.19683} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1198.268216] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b5fafe9-28f2-419a-831a-3b8c5e3e0623 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1198.268216] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-1b5fafe9-28f2-419a-831a-3b8c5e3e0623 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1198.268362] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7dc0ac71-59f9-4c12-b119-5e8f21bdd908 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.327391] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e21f186b-bd6d-4645-bde2-4c8882120cbb tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "a6a6c7de-656d-4358-977b-f62b471664da" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1198.327627] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e21f186b-bd6d-4645-bde2-4c8882120cbb tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "a6a6c7de-656d-4358-977b-f62b471664da" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1198.331981] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-1b5fafe9-28f2-419a-831a-3b8c5e3e0623 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1198.332216] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-1b5fafe9-28f2-419a-831a-3b8c5e3e0623 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Deleting contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1198.332463] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b5fafe9-28f2-419a-831a-3b8c5e3e0623 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Deleting the datastore file [datastore2] 28afc06e-f6e4-4044-8bfe-c93b1d9c1388 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1198.332759] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4d3bb183-27be-4ae8-8951-91fb1ccce670 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.338955] env[62385]: DEBUG oslo_vmware.api [None req-1b5fafe9-28f2-419a-831a-3b8c5e3e0623 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for the task: (returnval){ [ 1198.338955] env[62385]: value = "task-1206583" [ 1198.338955] env[62385]: _type = "Task" [ 1198.338955] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1198.347156] env[62385]: DEBUG oslo_vmware.api [None req-1b5fafe9-28f2-419a-831a-3b8c5e3e0623 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206583, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.831197] env[62385]: DEBUG nova.compute.utils [None req-e21f186b-bd6d-4645-bde2-4c8882120cbb tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1198.849699] env[62385]: DEBUG oslo_vmware.api [None req-1b5fafe9-28f2-419a-831a-3b8c5e3e0623 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Task: {'id': task-1206583, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.143753} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1198.849931] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b5fafe9-28f2-419a-831a-3b8c5e3e0623 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1198.850137] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-1b5fafe9-28f2-419a-831a-3b8c5e3e0623 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Deleted contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1198.850320] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-1b5fafe9-28f2-419a-831a-3b8c5e3e0623 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1198.850497] env[62385]: INFO nova.compute.manager [None req-1b5fafe9-28f2-419a-831a-3b8c5e3e0623 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1198.850730] env[62385]: DEBUG oslo.service.loopingcall [None req-1b5fafe9-28f2-419a-831a-3b8c5e3e0623 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1198.850921] env[62385]: DEBUG nova.compute.manager [-] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1198.851021] env[62385]: DEBUG nova.network.neutron [-] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1199.272342] env[62385]: DEBUG nova.compute.manager [req-cd6f46c4-8d89-4c71-b0a1-de74aabf0b05 req-f14f12fe-d9bd-4927-b61f-8e7f85007e30 service nova] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Received event network-vif-deleted-825ca180-3836-41d5-8d91-f55bc2ec9853 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1199.272718] env[62385]: INFO nova.compute.manager [req-cd6f46c4-8d89-4c71-b0a1-de74aabf0b05 req-f14f12fe-d9bd-4927-b61f-8e7f85007e30 service nova] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Neutron deleted interface 825ca180-3836-41d5-8d91-f55bc2ec9853; detaching it from the instance and deleting it from the info cache [ 1199.273846] env[62385]: DEBUG nova.network.neutron [req-cd6f46c4-8d89-4c71-b0a1-de74aabf0b05 req-f14f12fe-d9bd-4927-b61f-8e7f85007e30 service nova] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1199.334294] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e21f186b-bd6d-4645-bde2-4c8882120cbb tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "a6a6c7de-656d-4358-977b-f62b471664da" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1199.745055] env[62385]: DEBUG nova.network.neutron [-] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1199.775380] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7df3eaac-a261-48d1-b0ff-d25b1fd7bb3c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.785585] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5441b846-3e07-46d5-bc89-701853f5a16f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.811442] env[62385]: DEBUG nova.compute.manager [req-cd6f46c4-8d89-4c71-b0a1-de74aabf0b05 req-f14f12fe-d9bd-4927-b61f-8e7f85007e30 service nova] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Detach interface failed, port_id=825ca180-3836-41d5-8d91-f55bc2ec9853, reason: Instance 28afc06e-f6e4-4044-8bfe-c93b1d9c1388 could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1200.247168] env[62385]: INFO nova.compute.manager [-] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Took 1.40 seconds to deallocate network for instance. [ 1200.406828] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e21f186b-bd6d-4645-bde2-4c8882120cbb tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "a6a6c7de-656d-4358-977b-f62b471664da" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1200.407114] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e21f186b-bd6d-4645-bde2-4c8882120cbb tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "a6a6c7de-656d-4358-977b-f62b471664da" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1200.407366] env[62385]: INFO nova.compute.manager [None req-e21f186b-bd6d-4645-bde2-4c8882120cbb tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Attaching volume 7c2a17b9-4b5d-4085-9876-2a820028f37b to /dev/sdb [ 1200.437924] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2d87d17-f5ba-4e99-b881-0959cf0d6711 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.444888] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e68ad47b-faad-4458-92d7-f4a52e79b82b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.457921] env[62385]: DEBUG nova.virt.block_device [None req-e21f186b-bd6d-4645-bde2-4c8882120cbb tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Updating existing volume attachment record: 7b5e84c5-e6b9-4cde-8d19-1a0f25b0d174 {{(pid=62385) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1200.755238] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1b5fafe9-28f2-419a-831a-3b8c5e3e0623 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1200.755238] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1b5fafe9-28f2-419a-831a-3b8c5e3e0623 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1200.755238] env[62385]: DEBUG nova.objects.instance [None req-1b5fafe9-28f2-419a-831a-3b8c5e3e0623 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lazy-loading 'resources' on Instance uuid 28afc06e-f6e4-4044-8bfe-c93b1d9c1388 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1201.316821] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ac9a3c2-3a0b-41b6-a23f-a465cd81cd9a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.323953] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3a28518-4d71-4c2d-9147-175e213d3330 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.359846] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3c20d8d-6ba9-4db2-b5bc-ba6474670b03 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.367088] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c133673-b249-4632-8c12-5e5af4af2c62 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.381843] env[62385]: DEBUG nova.compute.provider_tree [None req-1b5fafe9-28f2-419a-831a-3b8c5e3e0623 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1201.884496] env[62385]: DEBUG nova.scheduler.client.report [None req-1b5fafe9-28f2-419a-831a-3b8c5e3e0623 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1202.389272] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1b5fafe9-28f2-419a-831a-3b8c5e3e0623 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.635s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1202.410398] env[62385]: INFO nova.scheduler.client.report [None req-1b5fafe9-28f2-419a-831a-3b8c5e3e0623 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Deleted allocations for instance 28afc06e-f6e4-4044-8bfe-c93b1d9c1388 [ 1202.918158] env[62385]: DEBUG oslo_concurrency.lockutils [None req-1b5fafe9-28f2-419a-831a-3b8c5e3e0623 tempest-ServerActionsTestJSON-1956545753 tempest-ServerActionsTestJSON-1956545753-project-member] Lock "28afc06e-f6e4-4044-8bfe-c93b1d9c1388" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.182s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1205.001481] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-e21f186b-bd6d-4645-bde2-4c8882120cbb tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Volume attach. Driver type: vmdk {{(pid=62385) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1205.001771] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-e21f186b-bd6d-4645-bde2-4c8882120cbb tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261327', 'volume_id': '7c2a17b9-4b5d-4085-9876-2a820028f37b', 'name': 'volume-7c2a17b9-4b5d-4085-9876-2a820028f37b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a6a6c7de-656d-4358-977b-f62b471664da', 'attached_at': '', 'detached_at': '', 'volume_id': '7c2a17b9-4b5d-4085-9876-2a820028f37b', 'serial': '7c2a17b9-4b5d-4085-9876-2a820028f37b'} {{(pid=62385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1205.002715] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4328a03-f4da-4124-b7dc-ddeafd5020d0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.019332] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3d1907e-6b0c-4b1e-98bc-cfd037d65161 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.043677] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-e21f186b-bd6d-4645-bde2-4c8882120cbb tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Reconfiguring VM instance instance-00000073 to attach disk [datastore1] volume-7c2a17b9-4b5d-4085-9876-2a820028f37b/volume-7c2a17b9-4b5d-4085-9876-2a820028f37b.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1205.043968] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-229cd933-4070-4bed-bb11-9aedc0642037 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.061573] env[62385]: DEBUG oslo_vmware.api [None req-e21f186b-bd6d-4645-bde2-4c8882120cbb tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1205.061573] env[62385]: value = "task-1206588" [ 1205.061573] env[62385]: _type = "Task" [ 1205.061573] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.069417] env[62385]: DEBUG oslo_vmware.api [None req-e21f186b-bd6d-4645-bde2-4c8882120cbb tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206588, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.572354] env[62385]: DEBUG oslo_vmware.api [None req-e21f186b-bd6d-4645-bde2-4c8882120cbb tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206588, 'name': ReconfigVM_Task, 'duration_secs': 0.326137} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1205.572597] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-e21f186b-bd6d-4645-bde2-4c8882120cbb tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Reconfigured VM instance instance-00000073 to attach disk [datastore1] volume-7c2a17b9-4b5d-4085-9876-2a820028f37b/volume-7c2a17b9-4b5d-4085-9876-2a820028f37b.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1205.577788] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-875b0ef5-3193-45ea-b6d2-4d7a0e822413 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.592377] env[62385]: DEBUG oslo_vmware.api [None req-e21f186b-bd6d-4645-bde2-4c8882120cbb tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1205.592377] env[62385]: value = "task-1206589" [ 1205.592377] env[62385]: _type = "Task" [ 1205.592377] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.600100] env[62385]: DEBUG oslo_vmware.api [None req-e21f186b-bd6d-4645-bde2-4c8882120cbb tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206589, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.103817] env[62385]: DEBUG oslo_vmware.api [None req-e21f186b-bd6d-4645-bde2-4c8882120cbb tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206589, 'name': ReconfigVM_Task, 'duration_secs': 0.139484} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.104155] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-e21f186b-bd6d-4645-bde2-4c8882120cbb tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261327', 'volume_id': '7c2a17b9-4b5d-4085-9876-2a820028f37b', 'name': 'volume-7c2a17b9-4b5d-4085-9876-2a820028f37b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a6a6c7de-656d-4358-977b-f62b471664da', 'attached_at': '', 'detached_at': '', 'volume_id': '7c2a17b9-4b5d-4085-9876-2a820028f37b', 'serial': '7c2a17b9-4b5d-4085-9876-2a820028f37b'} {{(pid=62385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1207.140256] env[62385]: DEBUG nova.objects.instance [None req-e21f186b-bd6d-4645-bde2-4c8882120cbb tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lazy-loading 'flavor' on Instance uuid a6a6c7de-656d-4358-977b-f62b471664da {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1207.646973] env[62385]: DEBUG oslo_concurrency.lockutils [None req-e21f186b-bd6d-4645-bde2-4c8882120cbb tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "a6a6c7de-656d-4358-977b-f62b471664da" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.240s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1208.949855] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "aa7fdda1-3441-4bc2-b2a3-c27ce91baab9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1208.949855] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "aa7fdda1-3441-4bc2-b2a3-c27ce91baab9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1209.451193] env[62385]: DEBUG nova.compute.manager [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1209.974523] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1209.974523] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1209.976082] env[62385]: INFO nova.compute.claims [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1211.030362] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9d0d68d-5f39-4b53-b39c-0cf276a7fbd2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.037885] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6e83bc9-314b-44de-ab8d-b8d2cbdca26f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.067557] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ca1f588-e6b6-4584-b6bd-1be887482a0b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.074078] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32e6d160-e4d1-454c-bcfe-e39b5064780a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.086422] env[62385]: DEBUG nova.compute.provider_tree [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1211.589468] env[62385]: DEBUG nova.scheduler.client.report [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1212.093734] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.119s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1212.094296] env[62385]: DEBUG nova.compute.manager [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1212.601983] env[62385]: DEBUG nova.compute.utils [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1212.605718] env[62385]: DEBUG nova.compute.manager [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1212.605892] env[62385]: DEBUG nova.network.neutron [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1212.658345] env[62385]: DEBUG nova.policy [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4a584f1cc579408eb589777251c6b07b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4831e1f029ed4ffc8ce5bfe7d3f38dd3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 1212.928018] env[62385]: DEBUG nova.network.neutron [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Successfully created port: 8054d150-69f7-421b-9434-47067fb778c6 {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1213.111669] env[62385]: DEBUG nova.compute.manager [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1214.122402] env[62385]: DEBUG nova.compute.manager [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1214.147658] env[62385]: DEBUG nova.virt.hardware [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1214.147900] env[62385]: DEBUG nova.virt.hardware [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1214.148157] env[62385]: DEBUG nova.virt.hardware [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1214.148270] env[62385]: DEBUG nova.virt.hardware [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1214.148423] env[62385]: DEBUG nova.virt.hardware [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1214.148575] env[62385]: DEBUG nova.virt.hardware [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1214.148791] env[62385]: DEBUG nova.virt.hardware [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1214.148983] env[62385]: DEBUG nova.virt.hardware [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1214.149203] env[62385]: DEBUG nova.virt.hardware [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1214.149377] env[62385]: DEBUG nova.virt.hardware [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1214.149555] env[62385]: DEBUG nova.virt.hardware [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1214.150428] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e6c1826-0274-40b0-a061-5171d4a1ecbc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.158218] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-092894ad-aefe-4d42-a215-ad52c6c7f23b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.286647] env[62385]: DEBUG nova.compute.manager [req-a4fdcfd3-2b92-4426-ad51-32a889e4eb27 req-60309e21-aa2d-4edf-a06d-071f41def689 service nova] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Received event network-vif-plugged-8054d150-69f7-421b-9434-47067fb778c6 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1214.286811] env[62385]: DEBUG oslo_concurrency.lockutils [req-a4fdcfd3-2b92-4426-ad51-32a889e4eb27 req-60309e21-aa2d-4edf-a06d-071f41def689 service nova] Acquiring lock "aa7fdda1-3441-4bc2-b2a3-c27ce91baab9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1214.287011] env[62385]: DEBUG oslo_concurrency.lockutils [req-a4fdcfd3-2b92-4426-ad51-32a889e4eb27 req-60309e21-aa2d-4edf-a06d-071f41def689 service nova] Lock "aa7fdda1-3441-4bc2-b2a3-c27ce91baab9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1214.287171] env[62385]: DEBUG oslo_concurrency.lockutils [req-a4fdcfd3-2b92-4426-ad51-32a889e4eb27 req-60309e21-aa2d-4edf-a06d-071f41def689 service nova] Lock "aa7fdda1-3441-4bc2-b2a3-c27ce91baab9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1214.287381] env[62385]: DEBUG nova.compute.manager [req-a4fdcfd3-2b92-4426-ad51-32a889e4eb27 req-60309e21-aa2d-4edf-a06d-071f41def689 service nova] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] No waiting events found dispatching network-vif-plugged-8054d150-69f7-421b-9434-47067fb778c6 {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1214.287539] env[62385]: WARNING nova.compute.manager [req-a4fdcfd3-2b92-4426-ad51-32a889e4eb27 req-60309e21-aa2d-4edf-a06d-071f41def689 service nova] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Received unexpected event network-vif-plugged-8054d150-69f7-421b-9434-47067fb778c6 for instance with vm_state building and task_state spawning. [ 1214.368493] env[62385]: DEBUG nova.network.neutron [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Successfully updated port: 8054d150-69f7-421b-9434-47067fb778c6 {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1214.871156] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "refresh_cache-aa7fdda1-3441-4bc2-b2a3-c27ce91baab9" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1214.871342] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquired lock "refresh_cache-aa7fdda1-3441-4bc2-b2a3-c27ce91baab9" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1214.871498] env[62385]: DEBUG nova.network.neutron [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1215.403646] env[62385]: DEBUG nova.network.neutron [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1215.519267] env[62385]: DEBUG nova.network.neutron [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Updating instance_info_cache with network_info: [{"id": "8054d150-69f7-421b-9434-47067fb778c6", "address": "fa:16:3e:60:6d:70", "network": {"id": "5aef0aec-abdf-4636-9d6b-ba476c7f0410", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1388512761-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4831e1f029ed4ffc8ce5bfe7d3f38dd3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "62f28d75-4e6a-4ae5-b8b3-d0652ea26d08", "external-id": "nsx-vlan-transportzone-801", "segmentation_id": 801, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8054d150-69", "ovs_interfaceid": "8054d150-69f7-421b-9434-47067fb778c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1216.021940] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Releasing lock "refresh_cache-aa7fdda1-3441-4bc2-b2a3-c27ce91baab9" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1216.022261] env[62385]: DEBUG nova.compute.manager [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Instance network_info: |[{"id": "8054d150-69f7-421b-9434-47067fb778c6", "address": "fa:16:3e:60:6d:70", "network": {"id": "5aef0aec-abdf-4636-9d6b-ba476c7f0410", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1388512761-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4831e1f029ed4ffc8ce5bfe7d3f38dd3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "62f28d75-4e6a-4ae5-b8b3-d0652ea26d08", "external-id": "nsx-vlan-transportzone-801", "segmentation_id": 801, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8054d150-69", "ovs_interfaceid": "8054d150-69f7-421b-9434-47067fb778c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1216.022729] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:60:6d:70', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '62f28d75-4e6a-4ae5-b8b3-d0652ea26d08', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8054d150-69f7-421b-9434-47067fb778c6', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1216.030297] env[62385]: DEBUG oslo.service.loopingcall [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1216.030501] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1216.030729] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d5fac941-2e1f-4cbb-9b8e-a813e7cff126 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.053221] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1216.053221] env[62385]: value = "task-1206590" [ 1216.053221] env[62385]: _type = "Task" [ 1216.053221] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1216.061650] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206590, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.311937] env[62385]: DEBUG nova.compute.manager [req-42e97854-d6f1-4760-92a6-e86733b7b1d4 req-4f1d4ba9-36db-416f-a225-3c3b4b03c2fe service nova] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Received event network-changed-8054d150-69f7-421b-9434-47067fb778c6 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1216.312163] env[62385]: DEBUG nova.compute.manager [req-42e97854-d6f1-4760-92a6-e86733b7b1d4 req-4f1d4ba9-36db-416f-a225-3c3b4b03c2fe service nova] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Refreshing instance network info cache due to event network-changed-8054d150-69f7-421b-9434-47067fb778c6. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1216.312509] env[62385]: DEBUG oslo_concurrency.lockutils [req-42e97854-d6f1-4760-92a6-e86733b7b1d4 req-4f1d4ba9-36db-416f-a225-3c3b4b03c2fe service nova] Acquiring lock "refresh_cache-aa7fdda1-3441-4bc2-b2a3-c27ce91baab9" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1216.312712] env[62385]: DEBUG oslo_concurrency.lockutils [req-42e97854-d6f1-4760-92a6-e86733b7b1d4 req-4f1d4ba9-36db-416f-a225-3c3b4b03c2fe service nova] Acquired lock "refresh_cache-aa7fdda1-3441-4bc2-b2a3-c27ce91baab9" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1216.312962] env[62385]: DEBUG nova.network.neutron [req-42e97854-d6f1-4760-92a6-e86733b7b1d4 req-4f1d4ba9-36db-416f-a225-3c3b4b03c2fe service nova] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Refreshing network info cache for port 8054d150-69f7-421b-9434-47067fb778c6 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1216.562968] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206590, 'name': CreateVM_Task, 'duration_secs': 0.31217} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1216.563330] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1216.563836] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1216.564017] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1216.564347] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1216.564594] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-738c55b1-a7d1-4198-90e1-9386bd2a4f00 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.568765] env[62385]: DEBUG oslo_vmware.api [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1216.568765] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c81848-4d39-cab4-2d9e-1bee4d0b0016" [ 1216.568765] env[62385]: _type = "Task" [ 1216.568765] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1216.575900] env[62385]: DEBUG oslo_vmware.api [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c81848-4d39-cab4-2d9e-1bee4d0b0016, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.995465] env[62385]: DEBUG nova.network.neutron [req-42e97854-d6f1-4760-92a6-e86733b7b1d4 req-4f1d4ba9-36db-416f-a225-3c3b4b03c2fe service nova] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Updated VIF entry in instance network info cache for port 8054d150-69f7-421b-9434-47067fb778c6. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1216.995825] env[62385]: DEBUG nova.network.neutron [req-42e97854-d6f1-4760-92a6-e86733b7b1d4 req-4f1d4ba9-36db-416f-a225-3c3b4b03c2fe service nova] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Updating instance_info_cache with network_info: [{"id": "8054d150-69f7-421b-9434-47067fb778c6", "address": "fa:16:3e:60:6d:70", "network": {"id": "5aef0aec-abdf-4636-9d6b-ba476c7f0410", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1388512761-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4831e1f029ed4ffc8ce5bfe7d3f38dd3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "62f28d75-4e6a-4ae5-b8b3-d0652ea26d08", "external-id": "nsx-vlan-transportzone-801", "segmentation_id": 801, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8054d150-69", "ovs_interfaceid": "8054d150-69f7-421b-9434-47067fb778c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1217.079183] env[62385]: DEBUG oslo_vmware.api [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52c81848-4d39-cab4-2d9e-1bee4d0b0016, 'name': SearchDatastore_Task, 'duration_secs': 0.009241} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1217.079489] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1217.079729] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1217.080053] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1217.080151] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1217.080305] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1217.080564] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-97105d9b-f851-4526-9b6d-d4ff8e015169 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.088022] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1217.088238] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1217.088921] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4e7aac8-1046-474a-9c8f-5c995a130159 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.093614] env[62385]: DEBUG oslo_vmware.api [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1217.093614] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52d594ea-aa6c-7034-277f-bf3e383b7efc" [ 1217.093614] env[62385]: _type = "Task" [ 1217.093614] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1217.100416] env[62385]: DEBUG oslo_vmware.api [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52d594ea-aa6c-7034-277f-bf3e383b7efc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.498943] env[62385]: DEBUG oslo_concurrency.lockutils [req-42e97854-d6f1-4760-92a6-e86733b7b1d4 req-4f1d4ba9-36db-416f-a225-3c3b4b03c2fe service nova] Releasing lock "refresh_cache-aa7fdda1-3441-4bc2-b2a3-c27ce91baab9" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1217.605410] env[62385]: DEBUG oslo_vmware.api [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52d594ea-aa6c-7034-277f-bf3e383b7efc, 'name': SearchDatastore_Task, 'duration_secs': 0.00758} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1217.606189] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0396863a-5ed3-4c41-94ca-ad0ef7ea51d2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.610990] env[62385]: DEBUG oslo_vmware.api [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1217.610990] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]525581ae-1998-f881-5af9-9057c5ee16a9" [ 1217.610990] env[62385]: _type = "Task" [ 1217.610990] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1217.617910] env[62385]: DEBUG oslo_vmware.api [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]525581ae-1998-f881-5af9-9057c5ee16a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.121628] env[62385]: DEBUG oslo_vmware.api [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]525581ae-1998-f881-5af9-9057c5ee16a9, 'name': SearchDatastore_Task, 'duration_secs': 0.008618} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1218.121877] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1218.122157] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] aa7fdda1-3441-4bc2-b2a3-c27ce91baab9/aa7fdda1-3441-4bc2-b2a3-c27ce91baab9.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1218.122416] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c996c30c-1e9a-402c-92af-0cc561b2e141 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.128850] env[62385]: DEBUG oslo_vmware.api [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1218.128850] env[62385]: value = "task-1206591" [ 1218.128850] env[62385]: _type = "Task" [ 1218.128850] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1218.136383] env[62385]: DEBUG oslo_vmware.api [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206591, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.638850] env[62385]: DEBUG oslo_vmware.api [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206591, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.411375} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1218.639257] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] aa7fdda1-3441-4bc2-b2a3-c27ce91baab9/aa7fdda1-3441-4bc2-b2a3-c27ce91baab9.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1218.639330] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1218.639580] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-af8b91d5-0950-4b0d-acd7-5b265978e9bb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.645894] env[62385]: DEBUG oslo_vmware.api [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1218.645894] env[62385]: value = "task-1206592" [ 1218.645894] env[62385]: _type = "Task" [ 1218.645894] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1218.652823] env[62385]: DEBUG oslo_vmware.api [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206592, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.155946] env[62385]: DEBUG oslo_vmware.api [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206592, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06409} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1219.156227] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1219.156965] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50298df6-a512-4e4a-8673-c96e5bfcd250 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.178180] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Reconfiguring VM instance instance-00000074 to attach disk [datastore1] aa7fdda1-3441-4bc2-b2a3-c27ce91baab9/aa7fdda1-3441-4bc2-b2a3-c27ce91baab9.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1219.179028] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-64d98692-e727-4d72-a03a-e4b2c83d4df2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.196884] env[62385]: DEBUG oslo_vmware.api [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1219.196884] env[62385]: value = "task-1206593" [ 1219.196884] env[62385]: _type = "Task" [ 1219.196884] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1219.204326] env[62385]: DEBUG oslo_vmware.api [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206593, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.707171] env[62385]: DEBUG oslo_vmware.api [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206593, 'name': ReconfigVM_Task, 'duration_secs': 0.263733} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1219.707551] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Reconfigured VM instance instance-00000074 to attach disk [datastore1] aa7fdda1-3441-4bc2-b2a3-c27ce91baab9/aa7fdda1-3441-4bc2-b2a3-c27ce91baab9.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1219.708171] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b9d25c90-30ba-4012-ace9-175c0719119e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.714579] env[62385]: DEBUG oslo_vmware.api [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1219.714579] env[62385]: value = "task-1206594" [ 1219.714579] env[62385]: _type = "Task" [ 1219.714579] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1219.722233] env[62385]: DEBUG oslo_vmware.api [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206594, 'name': Rename_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.224291] env[62385]: DEBUG oslo_vmware.api [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206594, 'name': Rename_Task, 'duration_secs': 0.148012} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1220.224544] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1220.224783] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e35852df-d52b-449b-9cca-f1de2cfa5bad {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.231088] env[62385]: DEBUG oslo_vmware.api [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1220.231088] env[62385]: value = "task-1206595" [ 1220.231088] env[62385]: _type = "Task" [ 1220.231088] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1220.238142] env[62385]: DEBUG oslo_vmware.api [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206595, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.741542] env[62385]: DEBUG oslo_vmware.api [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206595, 'name': PowerOnVM_Task, 'duration_secs': 0.414166} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1220.741913] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1220.742015] env[62385]: INFO nova.compute.manager [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Took 6.62 seconds to spawn the instance on the hypervisor. [ 1220.743355] env[62385]: DEBUG nova.compute.manager [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1220.743355] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8edc8cae-2e13-4dd3-86a0-7a05b60f372b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.260969] env[62385]: INFO nova.compute.manager [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Took 11.30 seconds to build instance. [ 1221.763504] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ac409f26-0f75-4ee0-82c0-e5c04d5752e3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "aa7fdda1-3441-4bc2-b2a3-c27ce91baab9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.814s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1221.953925] env[62385]: DEBUG nova.compute.manager [req-4ee0bb74-3340-4baa-9a75-e76c5796caf7 req-117e1ad8-d5d2-47af-8e13-1f0601b389da service nova] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Received event network-changed-8054d150-69f7-421b-9434-47067fb778c6 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1221.953925] env[62385]: DEBUG nova.compute.manager [req-4ee0bb74-3340-4baa-9a75-e76c5796caf7 req-117e1ad8-d5d2-47af-8e13-1f0601b389da service nova] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Refreshing instance network info cache due to event network-changed-8054d150-69f7-421b-9434-47067fb778c6. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1221.953925] env[62385]: DEBUG oslo_concurrency.lockutils [req-4ee0bb74-3340-4baa-9a75-e76c5796caf7 req-117e1ad8-d5d2-47af-8e13-1f0601b389da service nova] Acquiring lock "refresh_cache-aa7fdda1-3441-4bc2-b2a3-c27ce91baab9" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1221.953925] env[62385]: DEBUG oslo_concurrency.lockutils [req-4ee0bb74-3340-4baa-9a75-e76c5796caf7 req-117e1ad8-d5d2-47af-8e13-1f0601b389da service nova] Acquired lock "refresh_cache-aa7fdda1-3441-4bc2-b2a3-c27ce91baab9" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1221.953925] env[62385]: DEBUG nova.network.neutron [req-4ee0bb74-3340-4baa-9a75-e76c5796caf7 req-117e1ad8-d5d2-47af-8e13-1f0601b389da service nova] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Refreshing network info cache for port 8054d150-69f7-421b-9434-47067fb778c6 {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1222.742182] env[62385]: DEBUG nova.network.neutron [req-4ee0bb74-3340-4baa-9a75-e76c5796caf7 req-117e1ad8-d5d2-47af-8e13-1f0601b389da service nova] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Updated VIF entry in instance network info cache for port 8054d150-69f7-421b-9434-47067fb778c6. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1222.742539] env[62385]: DEBUG nova.network.neutron [req-4ee0bb74-3340-4baa-9a75-e76c5796caf7 req-117e1ad8-d5d2-47af-8e13-1f0601b389da service nova] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Updating instance_info_cache with network_info: [{"id": "8054d150-69f7-421b-9434-47067fb778c6", "address": "fa:16:3e:60:6d:70", "network": {"id": "5aef0aec-abdf-4636-9d6b-ba476c7f0410", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1388512761-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.188", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4831e1f029ed4ffc8ce5bfe7d3f38dd3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "62f28d75-4e6a-4ae5-b8b3-d0652ea26d08", "external-id": "nsx-vlan-transportzone-801", "segmentation_id": 801, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8054d150-69", "ovs_interfaceid": "8054d150-69f7-421b-9434-47067fb778c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1222.995786] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5b3f946a-5f5b-41a1-af06-316fbcfbe05d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquiring lock "4fe375be-ba60-4601-bd89-610e12742df3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1222.996170] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5b3f946a-5f5b-41a1-af06-316fbcfbe05d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lock "4fe375be-ba60-4601-bd89-610e12742df3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1222.996298] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5b3f946a-5f5b-41a1-af06-316fbcfbe05d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquiring lock "4fe375be-ba60-4601-bd89-610e12742df3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1222.996931] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5b3f946a-5f5b-41a1-af06-316fbcfbe05d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lock "4fe375be-ba60-4601-bd89-610e12742df3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1222.997141] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5b3f946a-5f5b-41a1-af06-316fbcfbe05d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lock "4fe375be-ba60-4601-bd89-610e12742df3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1222.999197] env[62385]: INFO nova.compute.manager [None req-5b3f946a-5f5b-41a1-af06-316fbcfbe05d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Terminating instance [ 1223.000894] env[62385]: DEBUG nova.compute.manager [None req-5b3f946a-5f5b-41a1-af06-316fbcfbe05d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1223.001114] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-5b3f946a-5f5b-41a1-af06-316fbcfbe05d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1223.001943] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1d7c0c0-0885-47c2-b113-30656a249791 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.009488] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b3f946a-5f5b-41a1-af06-316fbcfbe05d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1223.009717] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-daaeb095-2a6b-4f78-a2fc-22709f40eddd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.015716] env[62385]: DEBUG oslo_vmware.api [None req-5b3f946a-5f5b-41a1-af06-316fbcfbe05d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1223.015716] env[62385]: value = "task-1206596" [ 1223.015716] env[62385]: _type = "Task" [ 1223.015716] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.022778] env[62385]: DEBUG oslo_vmware.api [None req-5b3f946a-5f5b-41a1-af06-316fbcfbe05d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206596, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.246066] env[62385]: DEBUG oslo_concurrency.lockutils [req-4ee0bb74-3340-4baa-9a75-e76c5796caf7 req-117e1ad8-d5d2-47af-8e13-1f0601b389da service nova] Releasing lock "refresh_cache-aa7fdda1-3441-4bc2-b2a3-c27ce91baab9" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1223.526136] env[62385]: DEBUG oslo_vmware.api [None req-5b3f946a-5f5b-41a1-af06-316fbcfbe05d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206596, 'name': PowerOffVM_Task, 'duration_secs': 0.183198} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1223.526378] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b3f946a-5f5b-41a1-af06-316fbcfbe05d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1223.526516] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-5b3f946a-5f5b-41a1-af06-316fbcfbe05d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1223.527049] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-423a5ab7-e104-4e63-8861-9348f7a99256 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.584919] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-5b3f946a-5f5b-41a1-af06-316fbcfbe05d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1223.585266] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-5b3f946a-5f5b-41a1-af06-316fbcfbe05d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Deleting contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1223.585514] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b3f946a-5f5b-41a1-af06-316fbcfbe05d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Deleting the datastore file [datastore2] 4fe375be-ba60-4601-bd89-610e12742df3 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1223.585859] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b82dc20f-e9aa-407d-ac57-faa58f2f8670 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.591872] env[62385]: DEBUG oslo_vmware.api [None req-5b3f946a-5f5b-41a1-af06-316fbcfbe05d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for the task: (returnval){ [ 1223.591872] env[62385]: value = "task-1206598" [ 1223.591872] env[62385]: _type = "Task" [ 1223.591872] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.599663] env[62385]: DEBUG oslo_vmware.api [None req-5b3f946a-5f5b-41a1-af06-316fbcfbe05d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206598, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.101572] env[62385]: DEBUG oslo_vmware.api [None req-5b3f946a-5f5b-41a1-af06-316fbcfbe05d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Task: {'id': task-1206598, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14061} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.101959] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b3f946a-5f5b-41a1-af06-316fbcfbe05d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1224.102154] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-5b3f946a-5f5b-41a1-af06-316fbcfbe05d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Deleted contents of the VM from datastore datastore2 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1224.102343] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-5b3f946a-5f5b-41a1-af06-316fbcfbe05d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1224.102524] env[62385]: INFO nova.compute.manager [None req-5b3f946a-5f5b-41a1-af06-316fbcfbe05d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1224.102769] env[62385]: DEBUG oslo.service.loopingcall [None req-5b3f946a-5f5b-41a1-af06-316fbcfbe05d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1224.102982] env[62385]: DEBUG nova.compute.manager [-] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1224.103092] env[62385]: DEBUG nova.network.neutron [-] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1224.506871] env[62385]: DEBUG nova.compute.manager [req-f7a153d9-01e8-46ec-af47-155810549239 req-e063653a-0654-4c3a-aabf-46fd74907442 service nova] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Received event network-vif-deleted-1e16fa5f-3cab-4323-8637-7ff6dffce79b {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1224.507155] env[62385]: INFO nova.compute.manager [req-f7a153d9-01e8-46ec-af47-155810549239 req-e063653a-0654-4c3a-aabf-46fd74907442 service nova] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Neutron deleted interface 1e16fa5f-3cab-4323-8637-7ff6dffce79b; detaching it from the instance and deleting it from the info cache [ 1224.507337] env[62385]: DEBUG nova.network.neutron [req-f7a153d9-01e8-46ec-af47-155810549239 req-e063653a-0654-4c3a-aabf-46fd74907442 service nova] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1224.986122] env[62385]: DEBUG nova.network.neutron [-] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1225.009513] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5000a9be-e539-4319-bdd0-da9fddba09aa {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.021952] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-771382d1-f35d-442a-9d82-5ad25c67ff1f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.042582] env[62385]: DEBUG nova.compute.manager [req-f7a153d9-01e8-46ec-af47-155810549239 req-e063653a-0654-4c3a-aabf-46fd74907442 service nova] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Detach interface failed, port_id=1e16fa5f-3cab-4323-8637-7ff6dffce79b, reason: Instance 4fe375be-ba60-4601-bd89-610e12742df3 could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1225.489434] env[62385]: INFO nova.compute.manager [-] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Took 1.39 seconds to deallocate network for instance. [ 1225.996625] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5b3f946a-5f5b-41a1-af06-316fbcfbe05d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1225.996916] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5b3f946a-5f5b-41a1-af06-316fbcfbe05d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1225.997157] env[62385]: DEBUG nova.objects.instance [None req-5b3f946a-5f5b-41a1-af06-316fbcfbe05d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lazy-loading 'resources' on Instance uuid 4fe375be-ba60-4601-bd89-610e12742df3 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1226.550426] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc36c73d-d8db-4822-96c8-55a3f6ee862e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.557848] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1187f928-a19e-4206-b2a2-f3a6bfc53854 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.586200] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18713e12-28fc-45ed-a3da-ef6e4886a22d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.592568] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24e8703c-3be0-4098-88e1-357130f4c831 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.606048] env[62385]: DEBUG nova.compute.provider_tree [None req-5b3f946a-5f5b-41a1-af06-316fbcfbe05d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1227.108654] env[62385]: DEBUG nova.scheduler.client.report [None req-5b3f946a-5f5b-41a1-af06-316fbcfbe05d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1227.614019] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5b3f946a-5f5b-41a1-af06-316fbcfbe05d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.617s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1227.635047] env[62385]: INFO nova.scheduler.client.report [None req-5b3f946a-5f5b-41a1-af06-316fbcfbe05d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Deleted allocations for instance 4fe375be-ba60-4601-bd89-610e12742df3 [ 1228.143417] env[62385]: DEBUG oslo_concurrency.lockutils [None req-5b3f946a-5f5b-41a1-af06-316fbcfbe05d tempest-AttachVolumeShelveTestJSON-909248923 tempest-AttachVolumeShelveTestJSON-909248923-project-member] Lock "4fe375be-ba60-4601-bd89-610e12742df3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.147s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1243.030447] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1243.030726] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1244.026064] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1244.029625] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1244.029784] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Starting heal instance info cache {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1244.029890] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Rebuilding the list of instances to heal {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1244.561287] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquiring lock "refresh_cache-a6a6c7de-656d-4358-977b-f62b471664da" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1244.561521] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquired lock "refresh_cache-a6a6c7de-656d-4358-977b-f62b471664da" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1244.561592] env[62385]: DEBUG nova.network.neutron [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Forcefully refreshing network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1244.561741] env[62385]: DEBUG nova.objects.instance [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lazy-loading 'info_cache' on Instance uuid a6a6c7de-656d-4358-977b-f62b471664da {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1246.276648] env[62385]: DEBUG nova.network.neutron [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Updating instance_info_cache with network_info: [{"id": "d861965b-f04d-4e6d-a368-a4d06566f9ae", "address": "fa:16:3e:75:be:11", "network": {"id": "5aef0aec-abdf-4636-9d6b-ba476c7f0410", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1388512761-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.141", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4831e1f029ed4ffc8ce5bfe7d3f38dd3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "62f28d75-4e6a-4ae5-b8b3-d0652ea26d08", "external-id": "nsx-vlan-transportzone-801", "segmentation_id": 801, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd861965b-f0", "ovs_interfaceid": "d861965b-f04d-4e6d-a368-a4d06566f9ae", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1246.779614] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Releasing lock "refresh_cache-a6a6c7de-656d-4358-977b-f62b471664da" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1246.779831] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Updated the network info_cache for instance {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1246.780082] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1246.780224] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1246.780376] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1246.780502] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62385) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1248.030064] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1249.029696] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager.update_available_resource {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1249.533496] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1249.533852] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1249.533894] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1249.534054] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62385) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1249.535055] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb15afe5-0b23-4774-9317-be71c7ac7d4d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.543137] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7800c3d-7469-444f-bc84-b59c13b52112 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.557063] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-046bcf21-07a2-4072-ba16-07219d3a8383 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.562940] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5664b85a-2b7a-4165-ba8f-58bd887c02fe {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.590779] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181014MB free_disk=155GB free_vcpus=48 pci_devices=None {{(pid=62385) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1249.590924] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1249.591104] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1250.616433] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance a6a6c7de-656d-4358-977b-f62b471664da actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1250.616688] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance aa7fdda1-3441-4bc2-b2a3-c27ce91baab9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1250.616831] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=62385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1250.616979] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=62385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1250.653038] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-512c14a5-f6c8-4f23-ad90-b42015270394 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.659763] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66ab1e78-9055-43fa-b9ea-bb284d0394f2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.690156] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e2b4e5e-5c6e-4f4f-b384-104424d4fb39 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.696842] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-682428b6-d11e-4b8b-aa36-62d26d2292bc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.709323] env[62385]: DEBUG nova.compute.provider_tree [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1251.212493] env[62385]: DEBUG nova.scheduler.client.report [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1251.717466] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62385) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1251.717961] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.127s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1259.820216] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a72be91e-0ec3-47a8-a563-7d54e5e6be58 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "aa7fdda1-3441-4bc2-b2a3-c27ce91baab9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1259.820659] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a72be91e-0ec3-47a8-a563-7d54e5e6be58 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "aa7fdda1-3441-4bc2-b2a3-c27ce91baab9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1259.820735] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a72be91e-0ec3-47a8-a563-7d54e5e6be58 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "aa7fdda1-3441-4bc2-b2a3-c27ce91baab9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1259.820933] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a72be91e-0ec3-47a8-a563-7d54e5e6be58 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "aa7fdda1-3441-4bc2-b2a3-c27ce91baab9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1259.821128] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a72be91e-0ec3-47a8-a563-7d54e5e6be58 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "aa7fdda1-3441-4bc2-b2a3-c27ce91baab9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1259.824807] env[62385]: INFO nova.compute.manager [None req-a72be91e-0ec3-47a8-a563-7d54e5e6be58 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Terminating instance [ 1259.826665] env[62385]: DEBUG nova.compute.manager [None req-a72be91e-0ec3-47a8-a563-7d54e5e6be58 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1259.826858] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a72be91e-0ec3-47a8-a563-7d54e5e6be58 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1259.827718] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e86c3508-544d-4d70-b4d5-e8fe7be24700 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.835638] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a72be91e-0ec3-47a8-a563-7d54e5e6be58 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1259.835853] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f92bed91-1db5-4207-823e-1f96ee59d5db {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.842039] env[62385]: DEBUG oslo_vmware.api [None req-a72be91e-0ec3-47a8-a563-7d54e5e6be58 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1259.842039] env[62385]: value = "task-1206600" [ 1259.842039] env[62385]: _type = "Task" [ 1259.842039] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1259.849327] env[62385]: DEBUG oslo_vmware.api [None req-a72be91e-0ec3-47a8-a563-7d54e5e6be58 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206600, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.351667] env[62385]: DEBUG oslo_vmware.api [None req-a72be91e-0ec3-47a8-a563-7d54e5e6be58 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206600, 'name': PowerOffVM_Task, 'duration_secs': 0.169683} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1260.351923] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-a72be91e-0ec3-47a8-a563-7d54e5e6be58 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1260.352110] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a72be91e-0ec3-47a8-a563-7d54e5e6be58 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1260.352355] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a70bd552-f220-4621-adf1-22659e0c283d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.458596] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a72be91e-0ec3-47a8-a563-7d54e5e6be58 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1261.458976] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a72be91e-0ec3-47a8-a563-7d54e5e6be58 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1261.459046] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a72be91e-0ec3-47a8-a563-7d54e5e6be58 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Deleting the datastore file [datastore1] aa7fdda1-3441-4bc2-b2a3-c27ce91baab9 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1261.459289] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-043bc445-855e-47bd-8200-69ec3c4d507f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.466180] env[62385]: DEBUG oslo_vmware.api [None req-a72be91e-0ec3-47a8-a563-7d54e5e6be58 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1261.466180] env[62385]: value = "task-1206602" [ 1261.466180] env[62385]: _type = "Task" [ 1261.466180] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1261.474760] env[62385]: DEBUG oslo_vmware.api [None req-a72be91e-0ec3-47a8-a563-7d54e5e6be58 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206602, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.975401] env[62385]: DEBUG oslo_vmware.api [None req-a72be91e-0ec3-47a8-a563-7d54e5e6be58 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206602, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.143718} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1261.975653] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-a72be91e-0ec3-47a8-a563-7d54e5e6be58 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1261.975844] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a72be91e-0ec3-47a8-a563-7d54e5e6be58 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1261.976029] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-a72be91e-0ec3-47a8-a563-7d54e5e6be58 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1261.976215] env[62385]: INFO nova.compute.manager [None req-a72be91e-0ec3-47a8-a563-7d54e5e6be58 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Took 2.15 seconds to destroy the instance on the hypervisor. [ 1261.976456] env[62385]: DEBUG oslo.service.loopingcall [None req-a72be91e-0ec3-47a8-a563-7d54e5e6be58 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1261.976646] env[62385]: DEBUG nova.compute.manager [-] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1261.976739] env[62385]: DEBUG nova.network.neutron [-] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1262.418095] env[62385]: DEBUG nova.compute.manager [req-900e4eb0-d370-4eb2-a52b-b7e051de0b82 req-890927dd-f620-4348-90ba-5f1b0f1d6523 service nova] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Received event network-vif-deleted-8054d150-69f7-421b-9434-47067fb778c6 {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1262.418095] env[62385]: INFO nova.compute.manager [req-900e4eb0-d370-4eb2-a52b-b7e051de0b82 req-890927dd-f620-4348-90ba-5f1b0f1d6523 service nova] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Neutron deleted interface 8054d150-69f7-421b-9434-47067fb778c6; detaching it from the instance and deleting it from the info cache [ 1262.418095] env[62385]: DEBUG nova.network.neutron [req-900e4eb0-d370-4eb2-a52b-b7e051de0b82 req-890927dd-f620-4348-90ba-5f1b0f1d6523 service nova] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1262.892703] env[62385]: DEBUG nova.network.neutron [-] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1262.920274] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aa5cd9b9-d767-43da-b29b-3eafd6824e8f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.930356] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e5b1069-4479-478a-89f8-08fdc66376c9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.952781] env[62385]: DEBUG nova.compute.manager [req-900e4eb0-d370-4eb2-a52b-b7e051de0b82 req-890927dd-f620-4348-90ba-5f1b0f1d6523 service nova] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Detach interface failed, port_id=8054d150-69f7-421b-9434-47067fb778c6, reason: Instance aa7fdda1-3441-4bc2-b2a3-c27ce91baab9 could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1263.395479] env[62385]: INFO nova.compute.manager [-] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Took 1.42 seconds to deallocate network for instance. [ 1263.902156] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a72be91e-0ec3-47a8-a563-7d54e5e6be58 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1263.902492] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a72be91e-0ec3-47a8-a563-7d54e5e6be58 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1263.902677] env[62385]: DEBUG nova.objects.instance [None req-a72be91e-0ec3-47a8-a563-7d54e5e6be58 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lazy-loading 'resources' on Instance uuid aa7fdda1-3441-4bc2-b2a3-c27ce91baab9 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1264.448982] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-718fbf6b-fd45-4e81-ae00-81a5391d1a99 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.456897] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b89c88be-34bc-48d6-a242-a8d4fa0ed489 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.486706] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80e869cf-9e3b-46fb-a3da-3b18ad8436bb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.493386] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-054429c2-5b4e-4e8e-9d30-77ee8a0dcda4 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.506025] env[62385]: DEBUG nova.compute.provider_tree [None req-a72be91e-0ec3-47a8-a563-7d54e5e6be58 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1265.009634] env[62385]: DEBUG nova.scheduler.client.report [None req-a72be91e-0ec3-47a8-a563-7d54e5e6be58 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1265.515218] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a72be91e-0ec3-47a8-a563-7d54e5e6be58 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.613s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1265.540408] env[62385]: INFO nova.scheduler.client.report [None req-a72be91e-0ec3-47a8-a563-7d54e5e6be58 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Deleted allocations for instance aa7fdda1-3441-4bc2-b2a3-c27ce91baab9 [ 1266.048284] env[62385]: DEBUG oslo_concurrency.lockutils [None req-a72be91e-0ec3-47a8-a563-7d54e5e6be58 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "aa7fdda1-3441-4bc2-b2a3-c27ce91baab9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.228s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1266.972366] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0965f72f-0ecb-4f21-8818-ce0f45897bc5 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "a6a6c7de-656d-4358-977b-f62b471664da" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1266.972654] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0965f72f-0ecb-4f21-8818-ce0f45897bc5 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "a6a6c7de-656d-4358-977b-f62b471664da" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1267.475963] env[62385]: INFO nova.compute.manager [None req-0965f72f-0ecb-4f21-8818-ce0f45897bc5 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Detaching volume 7c2a17b9-4b5d-4085-9876-2a820028f37b [ 1267.508057] env[62385]: INFO nova.virt.block_device [None req-0965f72f-0ecb-4f21-8818-ce0f45897bc5 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Attempting to driver detach volume 7c2a17b9-4b5d-4085-9876-2a820028f37b from mountpoint /dev/sdb [ 1267.508316] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-0965f72f-0ecb-4f21-8818-ce0f45897bc5 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Volume detach. Driver type: vmdk {{(pid=62385) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1267.508509] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-0965f72f-0ecb-4f21-8818-ce0f45897bc5 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261327', 'volume_id': '7c2a17b9-4b5d-4085-9876-2a820028f37b', 'name': 'volume-7c2a17b9-4b5d-4085-9876-2a820028f37b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a6a6c7de-656d-4358-977b-f62b471664da', 'attached_at': '', 'detached_at': '', 'volume_id': '7c2a17b9-4b5d-4085-9876-2a820028f37b', 'serial': '7c2a17b9-4b5d-4085-9876-2a820028f37b'} {{(pid=62385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1267.509473] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c4cd2ad-1fa6-4ee7-955c-d3b7e62cb24a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.546425] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff1f756c-52e6-4ebd-bc03-67f5313131d5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.556732] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f522026b-fb6c-40c4-9e90-6724472546a1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.587951] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1285478-1484-462b-a9c0-fcde07f242b2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.610973] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-0965f72f-0ecb-4f21-8818-ce0f45897bc5 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] The volume has not been displaced from its original location: [datastore1] volume-7c2a17b9-4b5d-4085-9876-2a820028f37b/volume-7c2a17b9-4b5d-4085-9876-2a820028f37b.vmdk. No consolidation needed. {{(pid=62385) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1267.619994] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-0965f72f-0ecb-4f21-8818-ce0f45897bc5 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Reconfiguring VM instance instance-00000073 to detach disk 2001 {{(pid=62385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1267.620471] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d57c67d1-ed39-48c3-bb49-8293ea77fbae {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.648718] env[62385]: DEBUG oslo_vmware.api [None req-0965f72f-0ecb-4f21-8818-ce0f45897bc5 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1267.648718] env[62385]: value = "task-1206603" [ 1267.648718] env[62385]: _type = "Task" [ 1267.648718] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1267.660596] env[62385]: DEBUG oslo_vmware.api [None req-0965f72f-0ecb-4f21-8818-ce0f45897bc5 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206603, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1268.158918] env[62385]: DEBUG oslo_vmware.api [None req-0965f72f-0ecb-4f21-8818-ce0f45897bc5 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206603, 'name': ReconfigVM_Task, 'duration_secs': 0.262837} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1268.159259] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-0965f72f-0ecb-4f21-8818-ce0f45897bc5 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Reconfigured VM instance instance-00000073 to detach disk 2001 {{(pid=62385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1268.163831] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0b014fa1-4893-4957-ae20-5e6e42444d4a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.178757] env[62385]: DEBUG oslo_vmware.api [None req-0965f72f-0ecb-4f21-8818-ce0f45897bc5 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1268.178757] env[62385]: value = "task-1206604" [ 1268.178757] env[62385]: _type = "Task" [ 1268.178757] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1268.186796] env[62385]: DEBUG oslo_vmware.api [None req-0965f72f-0ecb-4f21-8818-ce0f45897bc5 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206604, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1268.689073] env[62385]: DEBUG oslo_vmware.api [None req-0965f72f-0ecb-4f21-8818-ce0f45897bc5 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206604, 'name': ReconfigVM_Task, 'duration_secs': 0.126826} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1268.689421] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-0965f72f-0ecb-4f21-8818-ce0f45897bc5 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261327', 'volume_id': '7c2a17b9-4b5d-4085-9876-2a820028f37b', 'name': 'volume-7c2a17b9-4b5d-4085-9876-2a820028f37b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a6a6c7de-656d-4358-977b-f62b471664da', 'attached_at': '', 'detached_at': '', 'volume_id': '7c2a17b9-4b5d-4085-9876-2a820028f37b', 'serial': '7c2a17b9-4b5d-4085-9876-2a820028f37b'} {{(pid=62385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1269.230999] env[62385]: DEBUG nova.objects.instance [None req-0965f72f-0ecb-4f21-8818-ce0f45897bc5 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lazy-loading 'flavor' on Instance uuid a6a6c7de-656d-4358-977b-f62b471664da {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1270.239015] env[62385]: DEBUG oslo_concurrency.lockutils [None req-0965f72f-0ecb-4f21-8818-ce0f45897bc5 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "a6a6c7de-656d-4358-977b-f62b471664da" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.266s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1271.212066] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c3f3659-b8d3-4b5a-88d9-accc4dfae68d tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "a6a6c7de-656d-4358-977b-f62b471664da" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1271.212370] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c3f3659-b8d3-4b5a-88d9-accc4dfae68d tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "a6a6c7de-656d-4358-977b-f62b471664da" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1271.212650] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c3f3659-b8d3-4b5a-88d9-accc4dfae68d tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "a6a6c7de-656d-4358-977b-f62b471664da-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1271.212888] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c3f3659-b8d3-4b5a-88d9-accc4dfae68d tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "a6a6c7de-656d-4358-977b-f62b471664da-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1271.213155] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c3f3659-b8d3-4b5a-88d9-accc4dfae68d tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "a6a6c7de-656d-4358-977b-f62b471664da-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1271.215427] env[62385]: INFO nova.compute.manager [None req-4c3f3659-b8d3-4b5a-88d9-accc4dfae68d tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Terminating instance [ 1271.217252] env[62385]: DEBUG nova.compute.manager [None req-4c3f3659-b8d3-4b5a-88d9-accc4dfae68d tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1271.217510] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4c3f3659-b8d3-4b5a-88d9-accc4dfae68d tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1271.218452] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e36b8c6e-8393-4376-aa43-4074b262b5d3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.226740] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c3f3659-b8d3-4b5a-88d9-accc4dfae68d tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1271.226992] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e9b79940-b54b-447d-a1e1-a50390cf3efa {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.233179] env[62385]: DEBUG oslo_vmware.api [None req-4c3f3659-b8d3-4b5a-88d9-accc4dfae68d tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1271.233179] env[62385]: value = "task-1206605" [ 1271.233179] env[62385]: _type = "Task" [ 1271.233179] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1271.240862] env[62385]: DEBUG oslo_vmware.api [None req-4c3f3659-b8d3-4b5a-88d9-accc4dfae68d tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206605, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1271.742646] env[62385]: DEBUG oslo_vmware.api [None req-4c3f3659-b8d3-4b5a-88d9-accc4dfae68d tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206605, 'name': PowerOffVM_Task, 'duration_secs': 0.198514} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1271.742920] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c3f3659-b8d3-4b5a-88d9-accc4dfae68d tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1271.743102] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4c3f3659-b8d3-4b5a-88d9-accc4dfae68d tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1271.743348] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fa053583-2e53-4ed5-9dfd-237bec4604be {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.977079] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4c3f3659-b8d3-4b5a-88d9-accc4dfae68d tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1272.977079] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4c3f3659-b8d3-4b5a-88d9-accc4dfae68d tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1272.977079] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c3f3659-b8d3-4b5a-88d9-accc4dfae68d tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Deleting the datastore file [datastore1] a6a6c7de-656d-4358-977b-f62b471664da {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1272.977079] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6341fb22-51d6-4203-babd-605014827eed {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.982832] env[62385]: DEBUG oslo_vmware.api [None req-4c3f3659-b8d3-4b5a-88d9-accc4dfae68d tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1272.982832] env[62385]: value = "task-1206607" [ 1272.982832] env[62385]: _type = "Task" [ 1272.982832] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1272.990237] env[62385]: DEBUG oslo_vmware.api [None req-4c3f3659-b8d3-4b5a-88d9-accc4dfae68d tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206607, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1273.493227] env[62385]: DEBUG oslo_vmware.api [None req-4c3f3659-b8d3-4b5a-88d9-accc4dfae68d tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206607, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.130501} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1273.493487] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c3f3659-b8d3-4b5a-88d9-accc4dfae68d tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1273.493789] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4c3f3659-b8d3-4b5a-88d9-accc4dfae68d tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1273.494021] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-4c3f3659-b8d3-4b5a-88d9-accc4dfae68d tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1273.494272] env[62385]: INFO nova.compute.manager [None req-4c3f3659-b8d3-4b5a-88d9-accc4dfae68d tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Took 2.28 seconds to destroy the instance on the hypervisor. [ 1273.494565] env[62385]: DEBUG oslo.service.loopingcall [None req-4c3f3659-b8d3-4b5a-88d9-accc4dfae68d tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1273.494812] env[62385]: DEBUG nova.compute.manager [-] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1273.494918] env[62385]: DEBUG nova.network.neutron [-] [instance: a6a6c7de-656d-4358-977b-f62b471664da] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1273.907447] env[62385]: DEBUG nova.compute.manager [req-24e6ed58-d6e6-4cea-abd3-08df6971464a req-fcd718b1-86d0-4c38-bd90-cffcba8bd09d service nova] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Received event network-vif-deleted-d861965b-f04d-4e6d-a368-a4d06566f9ae {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1273.907658] env[62385]: INFO nova.compute.manager [req-24e6ed58-d6e6-4cea-abd3-08df6971464a req-fcd718b1-86d0-4c38-bd90-cffcba8bd09d service nova] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Neutron deleted interface d861965b-f04d-4e6d-a368-a4d06566f9ae; detaching it from the instance and deleting it from the info cache [ 1273.907834] env[62385]: DEBUG nova.network.neutron [req-24e6ed58-d6e6-4cea-abd3-08df6971464a req-fcd718b1-86d0-4c38-bd90-cffcba8bd09d service nova] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1274.383128] env[62385]: DEBUG nova.network.neutron [-] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1274.410197] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-67709cfb-a40b-4c6d-947a-e8087b5f71d1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.420319] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a04ad0d4-cd58-4d2d-a482-ac8411312fc7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.442576] env[62385]: DEBUG nova.compute.manager [req-24e6ed58-d6e6-4cea-abd3-08df6971464a req-fcd718b1-86d0-4c38-bd90-cffcba8bd09d service nova] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Detach interface failed, port_id=d861965b-f04d-4e6d-a368-a4d06566f9ae, reason: Instance a6a6c7de-656d-4358-977b-f62b471664da could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1274.886106] env[62385]: INFO nova.compute.manager [-] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Took 1.39 seconds to deallocate network for instance. [ 1275.392582] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c3f3659-b8d3-4b5a-88d9-accc4dfae68d tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1275.393032] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c3f3659-b8d3-4b5a-88d9-accc4dfae68d tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1275.393107] env[62385]: DEBUG nova.objects.instance [None req-4c3f3659-b8d3-4b5a-88d9-accc4dfae68d tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lazy-loading 'resources' on Instance uuid a6a6c7de-656d-4358-977b-f62b471664da {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1275.925196] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c85a49ef-fba9-4686-8315-913ed5379724 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.933830] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1d7b10f-e815-4770-a85a-814dfa6851d3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.962413] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb10152f-faa4-4fa8-8657-0c6e85463795 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.969280] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cffee605-3e18-4a0c-a579-722fdcdfa026 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.981608] env[62385]: DEBUG nova.compute.provider_tree [None req-4c3f3659-b8d3-4b5a-88d9-accc4dfae68d tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1276.484939] env[62385]: DEBUG nova.scheduler.client.report [None req-4c3f3659-b8d3-4b5a-88d9-accc4dfae68d tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1276.990375] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c3f3659-b8d3-4b5a-88d9-accc4dfae68d tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.597s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1277.012014] env[62385]: INFO nova.scheduler.client.report [None req-4c3f3659-b8d3-4b5a-88d9-accc4dfae68d tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Deleted allocations for instance a6a6c7de-656d-4358-977b-f62b471664da [ 1277.520643] env[62385]: DEBUG oslo_concurrency.lockutils [None req-4c3f3659-b8d3-4b5a-88d9-accc4dfae68d tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "a6a6c7de-656d-4358-977b-f62b471664da" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.307s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1279.331028] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "a600e434-fb16-4881-acb9-e5a55c03f152" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1279.331317] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "a600e434-fb16-4881-acb9-e5a55c03f152" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1279.835060] env[62385]: DEBUG nova.compute.manager [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1280.353339] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1280.353636] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1280.355182] env[62385]: INFO nova.compute.claims [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1281.389642] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-868b6109-4788-4851-933d-68378dc70bf2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.397277] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2908e2a6-0f27-4408-a17c-ca7b6d4f317e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.425761] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a8466d3-684f-47ce-aa60-7e69052402db {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.432253] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-290237f4-9f27-4297-8b68-93b03dfedf49 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.445764] env[62385]: DEBUG nova.compute.provider_tree [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1281.948488] env[62385]: DEBUG nova.scheduler.client.report [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1282.453581] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.100s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1282.454162] env[62385]: DEBUG nova.compute.manager [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1282.959194] env[62385]: DEBUG nova.compute.utils [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1282.960628] env[62385]: DEBUG nova.compute.manager [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1282.960803] env[62385]: DEBUG nova.network.neutron [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1283.010066] env[62385]: DEBUG nova.policy [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4a584f1cc579408eb589777251c6b07b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4831e1f029ed4ffc8ce5bfe7d3f38dd3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 1283.282526] env[62385]: DEBUG nova.network.neutron [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Successfully created port: 3f0854df-43f9-4363-9108-e37e1bee96fb {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1283.465056] env[62385]: DEBUG nova.compute.manager [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1284.474274] env[62385]: DEBUG nova.compute.manager [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1284.499186] env[62385]: DEBUG nova.virt.hardware [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1284.499428] env[62385]: DEBUG nova.virt.hardware [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1284.499590] env[62385]: DEBUG nova.virt.hardware [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1284.499786] env[62385]: DEBUG nova.virt.hardware [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1284.499945] env[62385]: DEBUG nova.virt.hardware [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1284.500107] env[62385]: DEBUG nova.virt.hardware [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1284.500324] env[62385]: DEBUG nova.virt.hardware [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1284.500487] env[62385]: DEBUG nova.virt.hardware [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1284.500657] env[62385]: DEBUG nova.virt.hardware [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1284.500885] env[62385]: DEBUG nova.virt.hardware [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1284.500993] env[62385]: DEBUG nova.virt.hardware [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1284.501863] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fe7f24d-af20-45f1-b04b-5fe86c390a7c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.509940] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9473e1d-9415-4461-8bc6-06829bdc3c33 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.649569] env[62385]: DEBUG nova.compute.manager [req-41d6901f-e801-4ad0-9574-289682df10e7 req-c70e2509-ea44-4f42-84d7-56bc1c759810 service nova] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Received event network-vif-plugged-3f0854df-43f9-4363-9108-e37e1bee96fb {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1284.649805] env[62385]: DEBUG oslo_concurrency.lockutils [req-41d6901f-e801-4ad0-9574-289682df10e7 req-c70e2509-ea44-4f42-84d7-56bc1c759810 service nova] Acquiring lock "a600e434-fb16-4881-acb9-e5a55c03f152-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1284.650039] env[62385]: DEBUG oslo_concurrency.lockutils [req-41d6901f-e801-4ad0-9574-289682df10e7 req-c70e2509-ea44-4f42-84d7-56bc1c759810 service nova] Lock "a600e434-fb16-4881-acb9-e5a55c03f152-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1284.650228] env[62385]: DEBUG oslo_concurrency.lockutils [req-41d6901f-e801-4ad0-9574-289682df10e7 req-c70e2509-ea44-4f42-84d7-56bc1c759810 service nova] Lock "a600e434-fb16-4881-acb9-e5a55c03f152-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1284.650436] env[62385]: DEBUG nova.compute.manager [req-41d6901f-e801-4ad0-9574-289682df10e7 req-c70e2509-ea44-4f42-84d7-56bc1c759810 service nova] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] No waiting events found dispatching network-vif-plugged-3f0854df-43f9-4363-9108-e37e1bee96fb {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1284.650619] env[62385]: WARNING nova.compute.manager [req-41d6901f-e801-4ad0-9574-289682df10e7 req-c70e2509-ea44-4f42-84d7-56bc1c759810 service nova] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Received unexpected event network-vif-plugged-3f0854df-43f9-4363-9108-e37e1bee96fb for instance with vm_state building and task_state spawning. [ 1284.735134] env[62385]: DEBUG nova.network.neutron [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Successfully updated port: 3f0854df-43f9-4363-9108-e37e1bee96fb {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1285.238059] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "refresh_cache-a600e434-fb16-4881-acb9-e5a55c03f152" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1285.238233] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquired lock "refresh_cache-a600e434-fb16-4881-acb9-e5a55c03f152" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1285.238519] env[62385]: DEBUG nova.network.neutron [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1285.771066] env[62385]: DEBUG nova.network.neutron [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1285.889781] env[62385]: DEBUG nova.network.neutron [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Updating instance_info_cache with network_info: [{"id": "3f0854df-43f9-4363-9108-e37e1bee96fb", "address": "fa:16:3e:2b:cc:3d", "network": {"id": "5aef0aec-abdf-4636-9d6b-ba476c7f0410", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1388512761-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4831e1f029ed4ffc8ce5bfe7d3f38dd3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "62f28d75-4e6a-4ae5-b8b3-d0652ea26d08", "external-id": "nsx-vlan-transportzone-801", "segmentation_id": 801, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f0854df-43", "ovs_interfaceid": "3f0854df-43f9-4363-9108-e37e1bee96fb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1286.393083] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Releasing lock "refresh_cache-a600e434-fb16-4881-acb9-e5a55c03f152" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1286.393612] env[62385]: DEBUG nova.compute.manager [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Instance network_info: |[{"id": "3f0854df-43f9-4363-9108-e37e1bee96fb", "address": "fa:16:3e:2b:cc:3d", "network": {"id": "5aef0aec-abdf-4636-9d6b-ba476c7f0410", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1388512761-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4831e1f029ed4ffc8ce5bfe7d3f38dd3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "62f28d75-4e6a-4ae5-b8b3-d0652ea26d08", "external-id": "nsx-vlan-transportzone-801", "segmentation_id": 801, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f0854df-43", "ovs_interfaceid": "3f0854df-43f9-4363-9108-e37e1bee96fb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1286.394033] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2b:cc:3d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '62f28d75-4e6a-4ae5-b8b3-d0652ea26d08', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3f0854df-43f9-4363-9108-e37e1bee96fb', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1286.401424] env[62385]: DEBUG oslo.service.loopingcall [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1286.401656] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1286.401898] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1c26108d-af30-4ede-a24f-3c6c7139436d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.424984] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1286.424984] env[62385]: value = "task-1206608" [ 1286.424984] env[62385]: _type = "Task" [ 1286.424984] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1286.432511] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206608, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1286.679049] env[62385]: DEBUG nova.compute.manager [req-b0b68be1-f7d0-4c7b-a3e3-379d55818b6f req-afb96016-fea0-4d3c-bc24-84cd2e814403 service nova] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Received event network-changed-3f0854df-43f9-4363-9108-e37e1bee96fb {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1286.679260] env[62385]: DEBUG nova.compute.manager [req-b0b68be1-f7d0-4c7b-a3e3-379d55818b6f req-afb96016-fea0-4d3c-bc24-84cd2e814403 service nova] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Refreshing instance network info cache due to event network-changed-3f0854df-43f9-4363-9108-e37e1bee96fb. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1286.679483] env[62385]: DEBUG oslo_concurrency.lockutils [req-b0b68be1-f7d0-4c7b-a3e3-379d55818b6f req-afb96016-fea0-4d3c-bc24-84cd2e814403 service nova] Acquiring lock "refresh_cache-a600e434-fb16-4881-acb9-e5a55c03f152" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1286.679633] env[62385]: DEBUG oslo_concurrency.lockutils [req-b0b68be1-f7d0-4c7b-a3e3-379d55818b6f req-afb96016-fea0-4d3c-bc24-84cd2e814403 service nova] Acquired lock "refresh_cache-a600e434-fb16-4881-acb9-e5a55c03f152" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1286.679796] env[62385]: DEBUG nova.network.neutron [req-b0b68be1-f7d0-4c7b-a3e3-379d55818b6f req-afb96016-fea0-4d3c-bc24-84cd2e814403 service nova] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Refreshing network info cache for port 3f0854df-43f9-4363-9108-e37e1bee96fb {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1286.934643] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206608, 'name': CreateVM_Task, 'duration_secs': 0.287283} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1286.935038] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1286.935490] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1286.935662] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1286.936013] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1286.936288] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a82ec77-1612-4ccb-a27c-0938aab393d9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.940913] env[62385]: DEBUG oslo_vmware.api [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1286.940913] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a3c7dc-c022-3116-67c7-fa4628739f58" [ 1286.940913] env[62385]: _type = "Task" [ 1286.940913] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1286.948793] env[62385]: DEBUG oslo_vmware.api [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a3c7dc-c022-3116-67c7-fa4628739f58, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1287.387479] env[62385]: DEBUG nova.network.neutron [req-b0b68be1-f7d0-4c7b-a3e3-379d55818b6f req-afb96016-fea0-4d3c-bc24-84cd2e814403 service nova] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Updated VIF entry in instance network info cache for port 3f0854df-43f9-4363-9108-e37e1bee96fb. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1287.387912] env[62385]: DEBUG nova.network.neutron [req-b0b68be1-f7d0-4c7b-a3e3-379d55818b6f req-afb96016-fea0-4d3c-bc24-84cd2e814403 service nova] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Updating instance_info_cache with network_info: [{"id": "3f0854df-43f9-4363-9108-e37e1bee96fb", "address": "fa:16:3e:2b:cc:3d", "network": {"id": "5aef0aec-abdf-4636-9d6b-ba476c7f0410", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1388512761-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4831e1f029ed4ffc8ce5bfe7d3f38dd3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "62f28d75-4e6a-4ae5-b8b3-d0652ea26d08", "external-id": "nsx-vlan-transportzone-801", "segmentation_id": 801, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f0854df-43", "ovs_interfaceid": "3f0854df-43f9-4363-9108-e37e1bee96fb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1287.451090] env[62385]: DEBUG oslo_vmware.api [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52a3c7dc-c022-3116-67c7-fa4628739f58, 'name': SearchDatastore_Task, 'duration_secs': 0.01036} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1287.451397] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1287.451637] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1287.451902] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1287.452074] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1287.452262] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1287.452513] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0cd8b670-8571-4aeb-9b38-0e4118a5b64d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.460618] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1287.460762] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1287.461445] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b9122f8-78cf-46d0-9261-e7fd88f45ee6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.466425] env[62385]: DEBUG oslo_vmware.api [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1287.466425] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5204fc3d-6631-6f5c-8e36-c31d523b2189" [ 1287.466425] env[62385]: _type = "Task" [ 1287.466425] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1287.473441] env[62385]: DEBUG oslo_vmware.api [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5204fc3d-6631-6f5c-8e36-c31d523b2189, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1287.891081] env[62385]: DEBUG oslo_concurrency.lockutils [req-b0b68be1-f7d0-4c7b-a3e3-379d55818b6f req-afb96016-fea0-4d3c-bc24-84cd2e814403 service nova] Releasing lock "refresh_cache-a600e434-fb16-4881-acb9-e5a55c03f152" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1287.976589] env[62385]: DEBUG oslo_vmware.api [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5204fc3d-6631-6f5c-8e36-c31d523b2189, 'name': SearchDatastore_Task, 'duration_secs': 0.007911} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1287.977344] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7508e6ea-aea5-4d8b-bf86-49088feeaa7b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.982419] env[62385]: DEBUG oslo_vmware.api [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1287.982419] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5213171f-1cad-87a7-36d1-61d7590d4dc1" [ 1287.982419] env[62385]: _type = "Task" [ 1287.982419] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1287.989776] env[62385]: DEBUG oslo_vmware.api [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5213171f-1cad-87a7-36d1-61d7590d4dc1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1288.492580] env[62385]: DEBUG oslo_vmware.api [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5213171f-1cad-87a7-36d1-61d7590d4dc1, 'name': SearchDatastore_Task, 'duration_secs': 0.008499} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1288.492841] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1288.493111] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] a600e434-fb16-4881-acb9-e5a55c03f152/a600e434-fb16-4881-acb9-e5a55c03f152.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1288.493367] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0f3ffedb-1a35-4db2-92df-ded34ee019d5 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.500542] env[62385]: DEBUG oslo_vmware.api [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1288.500542] env[62385]: value = "task-1206609" [ 1288.500542] env[62385]: _type = "Task" [ 1288.500542] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1288.509902] env[62385]: DEBUG oslo_vmware.api [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206609, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1289.010248] env[62385]: DEBUG oslo_vmware.api [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206609, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.407412} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1289.010616] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] a600e434-fb16-4881-acb9-e5a55c03f152/a600e434-fb16-4881-acb9-e5a55c03f152.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1289.010694] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1289.010987] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-73f26347-a37c-42db-8d19-65493e4fd692 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.017572] env[62385]: DEBUG oslo_vmware.api [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1289.017572] env[62385]: value = "task-1206610" [ 1289.017572] env[62385]: _type = "Task" [ 1289.017572] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1289.024446] env[62385]: DEBUG oslo_vmware.api [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206610, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1289.527769] env[62385]: DEBUG oslo_vmware.api [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206610, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060142} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1289.528037] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1289.528772] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-931f2db9-501d-4598-ae88-1e2fc6d9b75d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.549841] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Reconfiguring VM instance instance-00000075 to attach disk [datastore1] a600e434-fb16-4881-acb9-e5a55c03f152/a600e434-fb16-4881-acb9-e5a55c03f152.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1289.550276] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6aa9f220-996b-4258-a779-ba7d14448338 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.569578] env[62385]: DEBUG oslo_vmware.api [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1289.569578] env[62385]: value = "task-1206611" [ 1289.569578] env[62385]: _type = "Task" [ 1289.569578] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1289.577388] env[62385]: DEBUG oslo_vmware.api [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206611, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1290.079784] env[62385]: DEBUG oslo_vmware.api [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206611, 'name': ReconfigVM_Task, 'duration_secs': 0.288853} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1290.080168] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Reconfigured VM instance instance-00000075 to attach disk [datastore1] a600e434-fb16-4881-acb9-e5a55c03f152/a600e434-fb16-4881-acb9-e5a55c03f152.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1290.080719] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-38e802e6-ddff-43bf-bad4-675cbad8f229 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.087513] env[62385]: DEBUG oslo_vmware.api [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1290.087513] env[62385]: value = "task-1206612" [ 1290.087513] env[62385]: _type = "Task" [ 1290.087513] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1290.094408] env[62385]: DEBUG oslo_vmware.api [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206612, 'name': Rename_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1290.598415] env[62385]: DEBUG oslo_vmware.api [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206612, 'name': Rename_Task, 'duration_secs': 0.138774} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1290.598701] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1290.598955] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a58dfaae-da1b-4e18-9710-51f87ba807de {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.605136] env[62385]: DEBUG oslo_vmware.api [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1290.605136] env[62385]: value = "task-1206613" [ 1290.605136] env[62385]: _type = "Task" [ 1290.605136] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1290.612365] env[62385]: DEBUG oslo_vmware.api [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206613, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1291.115771] env[62385]: DEBUG oslo_vmware.api [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206613, 'name': PowerOnVM_Task, 'duration_secs': 0.436621} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1291.116175] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1291.116369] env[62385]: INFO nova.compute.manager [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Took 6.64 seconds to spawn the instance on the hypervisor. [ 1291.116558] env[62385]: DEBUG nova.compute.manager [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1291.117317] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7611face-bec4-4db8-9352-3b45085f41de {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.635209] env[62385]: INFO nova.compute.manager [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Took 11.30 seconds to build instance. [ 1292.137459] env[62385]: DEBUG oslo_concurrency.lockutils [None req-2e2eec76-bd5e-4aa4-ad12-75ce6ca75962 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "a600e434-fb16-4881-acb9-e5a55c03f152" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.806s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1292.362113] env[62385]: DEBUG nova.compute.manager [req-95102aec-7b85-40ff-9f97-3d82a4149157 req-8dc1d9fd-5eac-4b62-b2d6-24944b30b646 service nova] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Received event network-changed-3f0854df-43f9-4363-9108-e37e1bee96fb {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1292.362324] env[62385]: DEBUG nova.compute.manager [req-95102aec-7b85-40ff-9f97-3d82a4149157 req-8dc1d9fd-5eac-4b62-b2d6-24944b30b646 service nova] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Refreshing instance network info cache due to event network-changed-3f0854df-43f9-4363-9108-e37e1bee96fb. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1292.362543] env[62385]: DEBUG oslo_concurrency.lockutils [req-95102aec-7b85-40ff-9f97-3d82a4149157 req-8dc1d9fd-5eac-4b62-b2d6-24944b30b646 service nova] Acquiring lock "refresh_cache-a600e434-fb16-4881-acb9-e5a55c03f152" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1292.362691] env[62385]: DEBUG oslo_concurrency.lockutils [req-95102aec-7b85-40ff-9f97-3d82a4149157 req-8dc1d9fd-5eac-4b62-b2d6-24944b30b646 service nova] Acquired lock "refresh_cache-a600e434-fb16-4881-acb9-e5a55c03f152" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1292.362856] env[62385]: DEBUG nova.network.neutron [req-95102aec-7b85-40ff-9f97-3d82a4149157 req-8dc1d9fd-5eac-4b62-b2d6-24944b30b646 service nova] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Refreshing network info cache for port 3f0854df-43f9-4363-9108-e37e1bee96fb {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1293.064333] env[62385]: DEBUG nova.network.neutron [req-95102aec-7b85-40ff-9f97-3d82a4149157 req-8dc1d9fd-5eac-4b62-b2d6-24944b30b646 service nova] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Updated VIF entry in instance network info cache for port 3f0854df-43f9-4363-9108-e37e1bee96fb. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1293.064683] env[62385]: DEBUG nova.network.neutron [req-95102aec-7b85-40ff-9f97-3d82a4149157 req-8dc1d9fd-5eac-4b62-b2d6-24944b30b646 service nova] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Updating instance_info_cache with network_info: [{"id": "3f0854df-43f9-4363-9108-e37e1bee96fb", "address": "fa:16:3e:2b:cc:3d", "network": {"id": "5aef0aec-abdf-4636-9d6b-ba476c7f0410", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1388512761-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4831e1f029ed4ffc8ce5bfe7d3f38dd3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "62f28d75-4e6a-4ae5-b8b3-d0652ea26d08", "external-id": "nsx-vlan-transportzone-801", "segmentation_id": 801, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f0854df-43", "ovs_interfaceid": "3f0854df-43f9-4363-9108-e37e1bee96fb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1293.567771] env[62385]: DEBUG oslo_concurrency.lockutils [req-95102aec-7b85-40ff-9f97-3d82a4149157 req-8dc1d9fd-5eac-4b62-b2d6-24944b30b646 service nova] Releasing lock "refresh_cache-a600e434-fb16-4881-acb9-e5a55c03f152" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1305.718696] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1305.719126] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1305.719126] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1306.026160] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1306.029738] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1306.029901] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Starting heal instance info cache {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1306.533391] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Didn't find any instances for network info cache update. {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1306.533645] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1306.533778] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62385) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1308.030710] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1309.030904] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1310.030777] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager.update_available_resource {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1310.534179] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1310.534551] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1310.534684] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1310.534820] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62385) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1310.535743] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeb9d274-8d72-4cf0-9b33-00d8c4639643 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.543929] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc229c70-4c99-48a6-b126-5134edc6c6c8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.557276] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1fcd294-c936-4599-84da-b53752c9b48b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.563105] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-082bf53b-311e-4caf-ab3f-78d8d5d0689e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.591740] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181183MB free_disk=155GB free_vcpus=48 pci_devices=None {{(pid=62385) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1310.591837] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1310.592034] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1311.616476] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance a600e434-fb16-4881-acb9-e5a55c03f152 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1311.616754] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=62385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1311.616829] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=62385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1311.632216] env[62385]: DEBUG nova.scheduler.client.report [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Refreshing inventories for resource provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1311.643783] env[62385]: DEBUG nova.scheduler.client.report [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Updating ProviderTree inventory for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1311.643982] env[62385]: DEBUG nova.compute.provider_tree [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Updating inventory in ProviderTree for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1311.653629] env[62385]: DEBUG nova.scheduler.client.report [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Refreshing aggregate associations for resource provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a, aggregates: None {{(pid=62385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1311.671113] env[62385]: DEBUG nova.scheduler.client.report [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Refreshing trait associations for resource provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE {{(pid=62385) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1311.767423] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2492450-ea12-40d7-bebf-24585de5c196 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.775230] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da7cbcd0-6a18-4301-97c8-06fb2973d868 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.803536] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5f35239-4648-4e02-886a-8fa721853434 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.810457] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da5b1f0a-d2ad-40a7-b81c-30fff64cebd3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.822841] env[62385]: DEBUG nova.compute.provider_tree [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1312.325672] env[62385]: DEBUG nova.scheduler.client.report [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1312.830844] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62385) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1312.831244] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.239s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1315.825753] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1329.883940] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9c57454f-07e7-44e5-b404-6f156ff57666 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "a600e434-fb16-4881-acb9-e5a55c03f152" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1329.884256] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9c57454f-07e7-44e5-b404-6f156ff57666 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "a600e434-fb16-4881-acb9-e5a55c03f152" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1330.387100] env[62385]: DEBUG nova.compute.utils [None req-9c57454f-07e7-44e5-b404-6f156ff57666 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1330.890651] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9c57454f-07e7-44e5-b404-6f156ff57666 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "a600e434-fb16-4881-acb9-e5a55c03f152" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1331.950753] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9c57454f-07e7-44e5-b404-6f156ff57666 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "a600e434-fb16-4881-acb9-e5a55c03f152" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1331.951183] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9c57454f-07e7-44e5-b404-6f156ff57666 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "a600e434-fb16-4881-acb9-e5a55c03f152" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1331.951307] env[62385]: INFO nova.compute.manager [None req-9c57454f-07e7-44e5-b404-6f156ff57666 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Attaching volume ba7a10fb-547f-4c11-8558-0337bea07090 to /dev/sdb [ 1331.981456] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f0e21d9-0f04-4afe-b5e4-1095fd766bb0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.988440] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0e4fd8d-44eb-4d9f-bb96-6b7d8c8e10c9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.000688] env[62385]: DEBUG nova.virt.block_device [None req-9c57454f-07e7-44e5-b404-6f156ff57666 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Updating existing volume attachment record: 972e9de2-a52c-463a-a451-01c10bcb3e9b {{(pid=62385) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1336.544402] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c57454f-07e7-44e5-b404-6f156ff57666 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Volume attach. Driver type: vmdk {{(pid=62385) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1336.544675] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c57454f-07e7-44e5-b404-6f156ff57666 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261330', 'volume_id': 'ba7a10fb-547f-4c11-8558-0337bea07090', 'name': 'volume-ba7a10fb-547f-4c11-8558-0337bea07090', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a600e434-fb16-4881-acb9-e5a55c03f152', 'attached_at': '', 'detached_at': '', 'volume_id': 'ba7a10fb-547f-4c11-8558-0337bea07090', 'serial': 'ba7a10fb-547f-4c11-8558-0337bea07090'} {{(pid=62385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1336.545583] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fb84324-9d33-471c-9e09-2fecfe9ab19e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.562911] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-451e4651-562f-45ff-84f4-985ef307d02b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.589914] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c57454f-07e7-44e5-b404-6f156ff57666 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Reconfiguring VM instance instance-00000075 to attach disk [datastore2] volume-ba7a10fb-547f-4c11-8558-0337bea07090/volume-ba7a10fb-547f-4c11-8558-0337bea07090.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1336.590181] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f20cce0c-2071-4b64-a096-d2d38595c2ed {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.609203] env[62385]: DEBUG oslo_vmware.api [None req-9c57454f-07e7-44e5-b404-6f156ff57666 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1336.609203] env[62385]: value = "task-1206616" [ 1336.609203] env[62385]: _type = "Task" [ 1336.609203] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1336.617362] env[62385]: DEBUG oslo_vmware.api [None req-9c57454f-07e7-44e5-b404-6f156ff57666 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206616, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1337.118244] env[62385]: DEBUG oslo_vmware.api [None req-9c57454f-07e7-44e5-b404-6f156ff57666 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206616, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1337.619168] env[62385]: DEBUG oslo_vmware.api [None req-9c57454f-07e7-44e5-b404-6f156ff57666 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206616, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1338.121172] env[62385]: DEBUG oslo_vmware.api [None req-9c57454f-07e7-44e5-b404-6f156ff57666 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206616, 'name': ReconfigVM_Task, 'duration_secs': 1.272149} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1338.121172] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c57454f-07e7-44e5-b404-6f156ff57666 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Reconfigured VM instance instance-00000075 to attach disk [datastore2] volume-ba7a10fb-547f-4c11-8558-0337bea07090/volume-ba7a10fb-547f-4c11-8558-0337bea07090.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1338.125113] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-336e0768-0b34-4e60-8635-c234bfa01d97 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.139865] env[62385]: DEBUG oslo_vmware.api [None req-9c57454f-07e7-44e5-b404-6f156ff57666 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1338.139865] env[62385]: value = "task-1206617" [ 1338.139865] env[62385]: _type = "Task" [ 1338.139865] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1338.147329] env[62385]: DEBUG oslo_vmware.api [None req-9c57454f-07e7-44e5-b404-6f156ff57666 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206617, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1338.649856] env[62385]: DEBUG oslo_vmware.api [None req-9c57454f-07e7-44e5-b404-6f156ff57666 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206617, 'name': ReconfigVM_Task, 'duration_secs': 0.132575} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1338.651043] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c57454f-07e7-44e5-b404-6f156ff57666 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261330', 'volume_id': 'ba7a10fb-547f-4c11-8558-0337bea07090', 'name': 'volume-ba7a10fb-547f-4c11-8558-0337bea07090', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a600e434-fb16-4881-acb9-e5a55c03f152', 'attached_at': '', 'detached_at': '', 'volume_id': 'ba7a10fb-547f-4c11-8558-0337bea07090', 'serial': 'ba7a10fb-547f-4c11-8558-0337bea07090'} {{(pid=62385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1339.685271] env[62385]: DEBUG nova.objects.instance [None req-9c57454f-07e7-44e5-b404-6f156ff57666 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lazy-loading 'flavor' on Instance uuid a600e434-fb16-4881-acb9-e5a55c03f152 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1340.190469] env[62385]: DEBUG oslo_concurrency.lockutils [None req-9c57454f-07e7-44e5-b404-6f156ff57666 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "a600e434-fb16-4881-acb9-e5a55c03f152" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.239s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1340.420633] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f0f1b864-5ac0-48e5-96c6-6fd69323fbae tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "a600e434-fb16-4881-acb9-e5a55c03f152" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1340.420921] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f0f1b864-5ac0-48e5-96c6-6fd69323fbae tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "a600e434-fb16-4881-acb9-e5a55c03f152" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1340.923542] env[62385]: INFO nova.compute.manager [None req-f0f1b864-5ac0-48e5-96c6-6fd69323fbae tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Detaching volume ba7a10fb-547f-4c11-8558-0337bea07090 [ 1340.953479] env[62385]: INFO nova.virt.block_device [None req-f0f1b864-5ac0-48e5-96c6-6fd69323fbae tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Attempting to driver detach volume ba7a10fb-547f-4c11-8558-0337bea07090 from mountpoint /dev/sdb [ 1340.953743] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0f1b864-5ac0-48e5-96c6-6fd69323fbae tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Volume detach. Driver type: vmdk {{(pid=62385) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1340.953930] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0f1b864-5ac0-48e5-96c6-6fd69323fbae tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261330', 'volume_id': 'ba7a10fb-547f-4c11-8558-0337bea07090', 'name': 'volume-ba7a10fb-547f-4c11-8558-0337bea07090', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a600e434-fb16-4881-acb9-e5a55c03f152', 'attached_at': '', 'detached_at': '', 'volume_id': 'ba7a10fb-547f-4c11-8558-0337bea07090', 'serial': 'ba7a10fb-547f-4c11-8558-0337bea07090'} {{(pid=62385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1340.954865] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d086e4eb-4303-423d-a19b-2c71b7658108 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.976806] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-185071ae-5f8a-42ef-9d73-5ca5818fc59e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.983123] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2f63c0f-7fe7-495f-884d-4a73ac30d22a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.002538] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d506e422-b142-41fc-b3ae-4128386a5363 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.015987] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0f1b864-5ac0-48e5-96c6-6fd69323fbae tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] The volume has not been displaced from its original location: [datastore2] volume-ba7a10fb-547f-4c11-8558-0337bea07090/volume-ba7a10fb-547f-4c11-8558-0337bea07090.vmdk. No consolidation needed. {{(pid=62385) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1341.021064] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0f1b864-5ac0-48e5-96c6-6fd69323fbae tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Reconfiguring VM instance instance-00000075 to detach disk 2001 {{(pid=62385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1341.021314] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-219d8a33-01a3-4d8d-a457-9007a521ab71 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.038281] env[62385]: DEBUG oslo_vmware.api [None req-f0f1b864-5ac0-48e5-96c6-6fd69323fbae tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1341.038281] env[62385]: value = "task-1206618" [ 1341.038281] env[62385]: _type = "Task" [ 1341.038281] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1341.045259] env[62385]: DEBUG oslo_vmware.api [None req-f0f1b864-5ac0-48e5-96c6-6fd69323fbae tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206618, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1341.547489] env[62385]: DEBUG oslo_vmware.api [None req-f0f1b864-5ac0-48e5-96c6-6fd69323fbae tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206618, 'name': ReconfigVM_Task, 'duration_secs': 0.212218} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1341.547766] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0f1b864-5ac0-48e5-96c6-6fd69323fbae tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Reconfigured VM instance instance-00000075 to detach disk 2001 {{(pid=62385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1341.552317] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e38da422-8181-485f-ab03-ac39de72254a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.566621] env[62385]: DEBUG oslo_vmware.api [None req-f0f1b864-5ac0-48e5-96c6-6fd69323fbae tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1341.566621] env[62385]: value = "task-1206619" [ 1341.566621] env[62385]: _type = "Task" [ 1341.566621] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1341.573872] env[62385]: DEBUG oslo_vmware.api [None req-f0f1b864-5ac0-48e5-96c6-6fd69323fbae tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206619, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1342.076087] env[62385]: DEBUG oslo_vmware.api [None req-f0f1b864-5ac0-48e5-96c6-6fd69323fbae tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206619, 'name': ReconfigVM_Task, 'duration_secs': 0.135999} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1342.076400] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0f1b864-5ac0-48e5-96c6-6fd69323fbae tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261330', 'volume_id': 'ba7a10fb-547f-4c11-8558-0337bea07090', 'name': 'volume-ba7a10fb-547f-4c11-8558-0337bea07090', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a600e434-fb16-4881-acb9-e5a55c03f152', 'attached_at': '', 'detached_at': '', 'volume_id': 'ba7a10fb-547f-4c11-8558-0337bea07090', 'serial': 'ba7a10fb-547f-4c11-8558-0337bea07090'} {{(pid=62385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1342.616343] env[62385]: DEBUG nova.objects.instance [None req-f0f1b864-5ac0-48e5-96c6-6fd69323fbae tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lazy-loading 'flavor' on Instance uuid a600e434-fb16-4881-acb9-e5a55c03f152 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1343.624102] env[62385]: DEBUG oslo_concurrency.lockutils [None req-f0f1b864-5ac0-48e5-96c6-6fd69323fbae tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "a600e434-fb16-4881-acb9-e5a55c03f152" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.203s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1344.667824] env[62385]: DEBUG oslo_concurrency.lockutils [None req-127a1704-290f-4071-990b-172aef7d07b4 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "a600e434-fb16-4881-acb9-e5a55c03f152" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1344.668226] env[62385]: DEBUG oslo_concurrency.lockutils [None req-127a1704-290f-4071-990b-172aef7d07b4 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "a600e434-fb16-4881-acb9-e5a55c03f152" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1344.668369] env[62385]: DEBUG oslo_concurrency.lockutils [None req-127a1704-290f-4071-990b-172aef7d07b4 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "a600e434-fb16-4881-acb9-e5a55c03f152-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1344.668516] env[62385]: DEBUG oslo_concurrency.lockutils [None req-127a1704-290f-4071-990b-172aef7d07b4 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "a600e434-fb16-4881-acb9-e5a55c03f152-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1344.668703] env[62385]: DEBUG oslo_concurrency.lockutils [None req-127a1704-290f-4071-990b-172aef7d07b4 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "a600e434-fb16-4881-acb9-e5a55c03f152-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1344.670802] env[62385]: INFO nova.compute.manager [None req-127a1704-290f-4071-990b-172aef7d07b4 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Terminating instance [ 1344.672566] env[62385]: DEBUG nova.compute.manager [None req-127a1704-290f-4071-990b-172aef7d07b4 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1344.672760] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-127a1704-290f-4071-990b-172aef7d07b4 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1344.673589] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a30e0d8d-e3f9-44cc-aaca-20a4c0c40319 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.681166] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-127a1704-290f-4071-990b-172aef7d07b4 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1344.681390] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ea583451-5361-4439-938e-0dc8973698f6 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.688031] env[62385]: DEBUG oslo_vmware.api [None req-127a1704-290f-4071-990b-172aef7d07b4 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1344.688031] env[62385]: value = "task-1206620" [ 1344.688031] env[62385]: _type = "Task" [ 1344.688031] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1344.695022] env[62385]: DEBUG oslo_vmware.api [None req-127a1704-290f-4071-990b-172aef7d07b4 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206620, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1345.198122] env[62385]: DEBUG oslo_vmware.api [None req-127a1704-290f-4071-990b-172aef7d07b4 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206620, 'name': PowerOffVM_Task, 'duration_secs': 0.171278} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1345.198418] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-127a1704-290f-4071-990b-172aef7d07b4 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1345.198618] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-127a1704-290f-4071-990b-172aef7d07b4 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1345.198869] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5405ba09-339d-4b96-8153-448ee3b3cd8a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.260839] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-127a1704-290f-4071-990b-172aef7d07b4 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1345.261122] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-127a1704-290f-4071-990b-172aef7d07b4 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1345.261288] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-127a1704-290f-4071-990b-172aef7d07b4 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Deleting the datastore file [datastore1] a600e434-fb16-4881-acb9-e5a55c03f152 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1345.261549] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2df26d09-4ea7-43b6-b898-e631a5cb942f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.269247] env[62385]: DEBUG oslo_vmware.api [None req-127a1704-290f-4071-990b-172aef7d07b4 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1345.269247] env[62385]: value = "task-1206622" [ 1345.269247] env[62385]: _type = "Task" [ 1345.269247] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1345.276475] env[62385]: DEBUG oslo_vmware.api [None req-127a1704-290f-4071-990b-172aef7d07b4 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206622, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1345.779263] env[62385]: DEBUG oslo_vmware.api [None req-127a1704-290f-4071-990b-172aef7d07b4 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206622, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.132733} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1345.779678] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-127a1704-290f-4071-990b-172aef7d07b4 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1345.779678] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-127a1704-290f-4071-990b-172aef7d07b4 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1345.779862] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-127a1704-290f-4071-990b-172aef7d07b4 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1345.780052] env[62385]: INFO nova.compute.manager [None req-127a1704-290f-4071-990b-172aef7d07b4 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1345.780303] env[62385]: DEBUG oslo.service.loopingcall [None req-127a1704-290f-4071-990b-172aef7d07b4 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1345.780496] env[62385]: DEBUG nova.compute.manager [-] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1345.780592] env[62385]: DEBUG nova.network.neutron [-] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1346.204704] env[62385]: DEBUG nova.compute.manager [req-c57a1217-62cc-4d9c-bc19-a49e8d350649 req-d02e8241-e9f8-43c2-b0be-371333eefbe0 service nova] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Received event network-vif-deleted-3f0854df-43f9-4363-9108-e37e1bee96fb {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1346.204854] env[62385]: INFO nova.compute.manager [req-c57a1217-62cc-4d9c-bc19-a49e8d350649 req-d02e8241-e9f8-43c2-b0be-371333eefbe0 service nova] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Neutron deleted interface 3f0854df-43f9-4363-9108-e37e1bee96fb; detaching it from the instance and deleting it from the info cache [ 1346.205034] env[62385]: DEBUG nova.network.neutron [req-c57a1217-62cc-4d9c-bc19-a49e8d350649 req-d02e8241-e9f8-43c2-b0be-371333eefbe0 service nova] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1346.683810] env[62385]: DEBUG nova.network.neutron [-] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1346.708014] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bed8a334-385f-477d-b613-899e16841a12 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.718129] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-012f5487-14da-4fd7-8fa6-5fb26130dc0b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.741977] env[62385]: DEBUG nova.compute.manager [req-c57a1217-62cc-4d9c-bc19-a49e8d350649 req-d02e8241-e9f8-43c2-b0be-371333eefbe0 service nova] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Detach interface failed, port_id=3f0854df-43f9-4363-9108-e37e1bee96fb, reason: Instance a600e434-fb16-4881-acb9-e5a55c03f152 could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1347.187121] env[62385]: INFO nova.compute.manager [-] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Took 1.41 seconds to deallocate network for instance. [ 1347.693925] env[62385]: DEBUG oslo_concurrency.lockutils [None req-127a1704-290f-4071-990b-172aef7d07b4 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1347.694215] env[62385]: DEBUG oslo_concurrency.lockutils [None req-127a1704-290f-4071-990b-172aef7d07b4 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1347.694484] env[62385]: DEBUG nova.objects.instance [None req-127a1704-290f-4071-990b-172aef7d07b4 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lazy-loading 'resources' on Instance uuid a600e434-fb16-4881-acb9-e5a55c03f152 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1348.227634] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eca8a8f2-a8e1-4fc9-8ce6-8a48e7e4278b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.235247] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd933915-13b2-4cc4-a44c-ae5575fa5be1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.264670] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65f2bfbe-cd3b-4331-bde3-c74eb012405b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.271649] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54588c54-72ad-431f-9513-4064476afd63 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.284283] env[62385]: DEBUG nova.compute.provider_tree [None req-127a1704-290f-4071-990b-172aef7d07b4 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1348.787709] env[62385]: DEBUG nova.scheduler.client.report [None req-127a1704-290f-4071-990b-172aef7d07b4 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1349.293413] env[62385]: DEBUG oslo_concurrency.lockutils [None req-127a1704-290f-4071-990b-172aef7d07b4 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.599s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1349.313868] env[62385]: INFO nova.scheduler.client.report [None req-127a1704-290f-4071-990b-172aef7d07b4 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Deleted allocations for instance a600e434-fb16-4881-acb9-e5a55c03f152 [ 1349.823101] env[62385]: DEBUG oslo_concurrency.lockutils [None req-127a1704-290f-4071-990b-172aef7d07b4 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "a600e434-fb16-4881-acb9-e5a55c03f152" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.155s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1351.477170] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "0793179a-ce76-4404-a036-a72b6dac82e9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1351.477496] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "0793179a-ce76-4404-a036-a72b6dac82e9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1351.979618] env[62385]: DEBUG nova.compute.manager [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Starting instance... {{(pid=62385) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1352.499976] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1352.500276] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1352.501803] env[62385]: INFO nova.compute.claims [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1353.536794] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-208b0d6e-c765-433f-a65c-1b77ff462b0d {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1353.544690] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-679bec26-6a4a-4698-9ce3-4e4aa17c3b6c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1353.574439] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65212e8a-3586-4bbf-b6ab-8067cee3963e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1353.581437] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43740eef-2a98-4e80-a518-3cf18b58a0a8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1353.594535] env[62385]: DEBUG nova.compute.provider_tree [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1354.097044] env[62385]: DEBUG nova.scheduler.client.report [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1354.602216] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.102s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1354.602711] env[62385]: DEBUG nova.compute.manager [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Start building networks asynchronously for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1355.108142] env[62385]: DEBUG nova.compute.utils [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1355.109533] env[62385]: DEBUG nova.compute.manager [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Allocating IP information in the background. {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1355.110245] env[62385]: DEBUG nova.network.neutron [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] allocate_for_instance() {{(pid=62385) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1355.159072] env[62385]: DEBUG nova.policy [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4a584f1cc579408eb589777251c6b07b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4831e1f029ed4ffc8ce5bfe7d3f38dd3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62385) authorize /opt/stack/nova/nova/policy.py:201}} [ 1355.436439] env[62385]: DEBUG nova.network.neutron [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Successfully created port: debdc145-742b-4723-b7a0-f34cfa0f749d {{(pid=62385) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1355.612881] env[62385]: DEBUG nova.compute.manager [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Start building block device mappings for instance. {{(pid=62385) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1356.623072] env[62385]: DEBUG nova.compute.manager [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Start spawning the instance on the hypervisor. {{(pid=62385) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1356.648690] env[62385]: DEBUG nova.virt.hardware [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T02:15:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T02:15:41Z,direct_url=,disk_format='vmdk',id=c0abbb8d-77e7-4f22-a256-0faf0a781109,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='2e3b444561ba4b12b2f10b6744f6aa39',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T02:15:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1356.648947] env[62385]: DEBUG nova.virt.hardware [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Flavor limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1356.649130] env[62385]: DEBUG nova.virt.hardware [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Image limits 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1356.649319] env[62385]: DEBUG nova.virt.hardware [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Flavor pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1356.649470] env[62385]: DEBUG nova.virt.hardware [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Image pref 0:0:0 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1356.649622] env[62385]: DEBUG nova.virt.hardware [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62385) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1356.649830] env[62385]: DEBUG nova.virt.hardware [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1356.649992] env[62385]: DEBUG nova.virt.hardware [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1356.650177] env[62385]: DEBUG nova.virt.hardware [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Got 1 possible topologies {{(pid=62385) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1356.650399] env[62385]: DEBUG nova.virt.hardware [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1356.650512] env[62385]: DEBUG nova.virt.hardware [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62385) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1356.651393] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e545b42-7432-45ff-babb-7cca12fb6e34 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.659670] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea57caf9-1792-4c22-9a41-10e7c6d45846 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.866612] env[62385]: DEBUG nova.compute.manager [req-b8b5878c-1390-41f2-971d-137e3743e2b6 req-835b6628-f461-4513-a70c-8ae66a9ec9ef service nova] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Received event network-vif-plugged-debdc145-742b-4723-b7a0-f34cfa0f749d {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1356.866875] env[62385]: DEBUG oslo_concurrency.lockutils [req-b8b5878c-1390-41f2-971d-137e3743e2b6 req-835b6628-f461-4513-a70c-8ae66a9ec9ef service nova] Acquiring lock "0793179a-ce76-4404-a036-a72b6dac82e9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1356.867060] env[62385]: DEBUG oslo_concurrency.lockutils [req-b8b5878c-1390-41f2-971d-137e3743e2b6 req-835b6628-f461-4513-a70c-8ae66a9ec9ef service nova] Lock "0793179a-ce76-4404-a036-a72b6dac82e9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1356.867242] env[62385]: DEBUG oslo_concurrency.lockutils [req-b8b5878c-1390-41f2-971d-137e3743e2b6 req-835b6628-f461-4513-a70c-8ae66a9ec9ef service nova] Lock "0793179a-ce76-4404-a036-a72b6dac82e9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1356.867432] env[62385]: DEBUG nova.compute.manager [req-b8b5878c-1390-41f2-971d-137e3743e2b6 req-835b6628-f461-4513-a70c-8ae66a9ec9ef service nova] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] No waiting events found dispatching network-vif-plugged-debdc145-742b-4723-b7a0-f34cfa0f749d {{(pid=62385) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1356.867616] env[62385]: WARNING nova.compute.manager [req-b8b5878c-1390-41f2-971d-137e3743e2b6 req-835b6628-f461-4513-a70c-8ae66a9ec9ef service nova] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Received unexpected event network-vif-plugged-debdc145-742b-4723-b7a0-f34cfa0f749d for instance with vm_state building and task_state spawning. [ 1357.368598] env[62385]: DEBUG nova.network.neutron [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Successfully updated port: debdc145-742b-4723-b7a0-f34cfa0f749d {{(pid=62385) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1357.393290] env[62385]: DEBUG nova.compute.manager [req-709b772b-b396-44b4-8992-3d72a27cbfdc req-a7e46dec-2de1-43a5-8299-fdd4fb6e823f service nova] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Received event network-changed-debdc145-742b-4723-b7a0-f34cfa0f749d {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1357.393558] env[62385]: DEBUG nova.compute.manager [req-709b772b-b396-44b4-8992-3d72a27cbfdc req-a7e46dec-2de1-43a5-8299-fdd4fb6e823f service nova] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Refreshing instance network info cache due to event network-changed-debdc145-742b-4723-b7a0-f34cfa0f749d. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1357.393935] env[62385]: DEBUG oslo_concurrency.lockutils [req-709b772b-b396-44b4-8992-3d72a27cbfdc req-a7e46dec-2de1-43a5-8299-fdd4fb6e823f service nova] Acquiring lock "refresh_cache-0793179a-ce76-4404-a036-a72b6dac82e9" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1357.394190] env[62385]: DEBUG oslo_concurrency.lockutils [req-709b772b-b396-44b4-8992-3d72a27cbfdc req-a7e46dec-2de1-43a5-8299-fdd4fb6e823f service nova] Acquired lock "refresh_cache-0793179a-ce76-4404-a036-a72b6dac82e9" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1357.394534] env[62385]: DEBUG nova.network.neutron [req-709b772b-b396-44b4-8992-3d72a27cbfdc req-a7e46dec-2de1-43a5-8299-fdd4fb6e823f service nova] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Refreshing network info cache for port debdc145-742b-4723-b7a0-f34cfa0f749d {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1357.874927] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "refresh_cache-0793179a-ce76-4404-a036-a72b6dac82e9" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1357.927661] env[62385]: DEBUG nova.network.neutron [req-709b772b-b396-44b4-8992-3d72a27cbfdc req-a7e46dec-2de1-43a5-8299-fdd4fb6e823f service nova] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1357.996049] env[62385]: DEBUG nova.network.neutron [req-709b772b-b396-44b4-8992-3d72a27cbfdc req-a7e46dec-2de1-43a5-8299-fdd4fb6e823f service nova] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1358.499059] env[62385]: DEBUG oslo_concurrency.lockutils [req-709b772b-b396-44b4-8992-3d72a27cbfdc req-a7e46dec-2de1-43a5-8299-fdd4fb6e823f service nova] Releasing lock "refresh_cache-0793179a-ce76-4404-a036-a72b6dac82e9" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1358.499459] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquired lock "refresh_cache-0793179a-ce76-4404-a036-a72b6dac82e9" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1358.499616] env[62385]: DEBUG nova.network.neutron [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Building network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1359.028884] env[62385]: DEBUG nova.network.neutron [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Instance cache missing network info. {{(pid=62385) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1359.146420] env[62385]: DEBUG nova.network.neutron [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Updating instance_info_cache with network_info: [{"id": "debdc145-742b-4723-b7a0-f34cfa0f749d", "address": "fa:16:3e:58:ee:91", "network": {"id": "5aef0aec-abdf-4636-9d6b-ba476c7f0410", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1388512761-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4831e1f029ed4ffc8ce5bfe7d3f38dd3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "62f28d75-4e6a-4ae5-b8b3-d0652ea26d08", "external-id": "nsx-vlan-transportzone-801", "segmentation_id": 801, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdebdc145-74", "ovs_interfaceid": "debdc145-742b-4723-b7a0-f34cfa0f749d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1359.649091] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Releasing lock "refresh_cache-0793179a-ce76-4404-a036-a72b6dac82e9" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1359.649421] env[62385]: DEBUG nova.compute.manager [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Instance network_info: |[{"id": "debdc145-742b-4723-b7a0-f34cfa0f749d", "address": "fa:16:3e:58:ee:91", "network": {"id": "5aef0aec-abdf-4636-9d6b-ba476c7f0410", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1388512761-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4831e1f029ed4ffc8ce5bfe7d3f38dd3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "62f28d75-4e6a-4ae5-b8b3-d0652ea26d08", "external-id": "nsx-vlan-transportzone-801", "segmentation_id": 801, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdebdc145-74", "ovs_interfaceid": "debdc145-742b-4723-b7a0-f34cfa0f749d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62385) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1359.649887] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:58:ee:91', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '62f28d75-4e6a-4ae5-b8b3-d0652ea26d08', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'debdc145-742b-4723-b7a0-f34cfa0f749d', 'vif_model': 'vmxnet3'}] {{(pid=62385) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1359.658161] env[62385]: DEBUG oslo.service.loopingcall [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1359.658368] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Creating VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1359.658601] env[62385]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b9f089b6-991d-489b-8c34-f50680845ef1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.678210] env[62385]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1359.678210] env[62385]: value = "task-1206623" [ 1359.678210] env[62385]: _type = "Task" [ 1359.678210] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1359.685310] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206623, 'name': CreateVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1360.187890] env[62385]: DEBUG oslo_vmware.api [-] Task: {'id': task-1206623, 'name': CreateVM_Task, 'duration_secs': 0.298387} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1360.188332] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Created VM on the ESX host {{(pid=62385) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1360.195122] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1360.195320] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1360.195639] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1360.195885] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06bf9aa7-690d-4c20-8825-4cea45aeb8bb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.200253] env[62385]: DEBUG oslo_vmware.api [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1360.200253] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52747afa-200e-1314-1d37-5b05f5301d67" [ 1360.200253] env[62385]: _type = "Task" [ 1360.200253] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1360.207649] env[62385]: DEBUG oslo_vmware.api [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52747afa-200e-1314-1d37-5b05f5301d67, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1360.710694] env[62385]: DEBUG oslo_vmware.api [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52747afa-200e-1314-1d37-5b05f5301d67, 'name': SearchDatastore_Task, 'duration_secs': 0.010936} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1360.710998] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1360.711257] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Processing image c0abbb8d-77e7-4f22-a256-0faf0a781109 {{(pid=62385) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1360.711480] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1360.711629] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1360.711813] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1360.712076] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e1e583f3-8e66-448c-86a7-619ab76ad9db {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.719489] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62385) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1360.719660] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62385) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1360.720328] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d23cee1-e80f-4151-a41e-1c45a39c5500 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.724864] env[62385]: DEBUG oslo_vmware.api [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1360.724864] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52118f83-6e91-7e8d-4dce-e243b6480e6c" [ 1360.724864] env[62385]: _type = "Task" [ 1360.724864] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1360.731841] env[62385]: DEBUG oslo_vmware.api [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52118f83-6e91-7e8d-4dce-e243b6480e6c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1361.234914] env[62385]: DEBUG oslo_vmware.api [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]52118f83-6e91-7e8d-4dce-e243b6480e6c, 'name': SearchDatastore_Task, 'duration_secs': 0.008329} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1361.235621] env[62385]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8735ee1d-9a2e-4a7b-a5a4-b4baf3c6b5d8 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1361.240313] env[62385]: DEBUG oslo_vmware.api [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1361.240313] env[62385]: value = "session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5290dc86-6ec9-c2fc-4808-f3ea49735201" [ 1361.240313] env[62385]: _type = "Task" [ 1361.240313] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1361.247550] env[62385]: DEBUG oslo_vmware.api [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5290dc86-6ec9-c2fc-4808-f3ea49735201, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1361.751171] env[62385]: DEBUG oslo_vmware.api [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': session[52a80ddf-aeef-e5e1-f01e-39cbd9da09ac]5290dc86-6ec9-c2fc-4808-f3ea49735201, 'name': SearchDatastore_Task, 'duration_secs': 0.00861} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1361.751449] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1361.751718] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 0793179a-ce76-4404-a036-a72b6dac82e9/0793179a-ce76-4404-a036-a72b6dac82e9.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1361.751984] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4c010518-aa46-4957-aea4-280dd764fa7f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1361.758196] env[62385]: DEBUG oslo_vmware.api [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1361.758196] env[62385]: value = "task-1206624" [ 1361.758196] env[62385]: _type = "Task" [ 1361.758196] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1361.766261] env[62385]: DEBUG oslo_vmware.api [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206624, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1362.268052] env[62385]: DEBUG oslo_vmware.api [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206624, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.508316} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1362.268430] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c0abbb8d-77e7-4f22-a256-0faf0a781109/c0abbb8d-77e7-4f22-a256-0faf0a781109.vmdk to [datastore1] 0793179a-ce76-4404-a036-a72b6dac82e9/0793179a-ce76-4404-a036-a72b6dac82e9.vmdk {{(pid=62385) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1362.268511] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Extending root virtual disk to 1048576 {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1362.268733] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d5d5abb6-bdc2-4886-bccf-b13a189716a0 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.275521] env[62385]: DEBUG oslo_vmware.api [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1362.275521] env[62385]: value = "task-1206625" [ 1362.275521] env[62385]: _type = "Task" [ 1362.275521] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1362.282342] env[62385]: DEBUG oslo_vmware.api [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206625, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1362.785734] env[62385]: DEBUG oslo_vmware.api [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206625, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08101} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1362.786033] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Extended root virtual disk {{(pid=62385) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1362.786755] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-806a25df-7fa5-4219-b554-6d8a1d6a09dd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.807578] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Reconfiguring VM instance instance-00000076 to attach disk [datastore1] 0793179a-ce76-4404-a036-a72b6dac82e9/0793179a-ce76-4404-a036-a72b6dac82e9.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1362.807818] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-38fa1162-d8fe-4c74-8c52-1018b267704a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.826898] env[62385]: DEBUG oslo_vmware.api [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1362.826898] env[62385]: value = "task-1206626" [ 1362.826898] env[62385]: _type = "Task" [ 1362.826898] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1362.834750] env[62385]: DEBUG oslo_vmware.api [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206626, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1363.029615] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1363.336998] env[62385]: DEBUG oslo_vmware.api [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206626, 'name': ReconfigVM_Task, 'duration_secs': 0.254946} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1363.337468] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Reconfigured VM instance instance-00000076 to attach disk [datastore1] 0793179a-ce76-4404-a036-a72b6dac82e9/0793179a-ce76-4404-a036-a72b6dac82e9.vmdk or device None with type sparse {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1363.337919] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-339acd7c-0183-4569-b19f-18d4a8e10518 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1363.344137] env[62385]: DEBUG oslo_vmware.api [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1363.344137] env[62385]: value = "task-1206627" [ 1363.344137] env[62385]: _type = "Task" [ 1363.344137] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1363.353179] env[62385]: DEBUG oslo_vmware.api [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206627, 'name': Rename_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1363.854289] env[62385]: DEBUG oslo_vmware.api [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206627, 'name': Rename_Task, 'duration_secs': 0.134522} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1363.854700] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Powering on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1363.855042] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d1e7bda0-c077-4c3e-8369-9ea286900dfb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1363.860937] env[62385]: DEBUG oslo_vmware.api [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1363.860937] env[62385]: value = "task-1206628" [ 1363.860937] env[62385]: _type = "Task" [ 1363.860937] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1363.868286] env[62385]: DEBUG oslo_vmware.api [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206628, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1364.030134] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1364.370890] env[62385]: DEBUG oslo_vmware.api [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206628, 'name': PowerOnVM_Task, 'duration_secs': 0.432248} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1364.371226] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Powered on the VM {{(pid=62385) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1364.371376] env[62385]: INFO nova.compute.manager [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Took 7.75 seconds to spawn the instance on the hypervisor. [ 1364.371557] env[62385]: DEBUG nova.compute.manager [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Checking state {{(pid=62385) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1364.372330] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20bd8101-4941-4a11-9343-30aea931c214 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.890767] env[62385]: INFO nova.compute.manager [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Took 12.41 seconds to build instance. [ 1365.392862] env[62385]: DEBUG oslo_concurrency.lockutils [None req-b53de063-50bc-4bde-a8df-e2e2a82b07f0 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "0793179a-ce76-4404-a036-a72b6dac82e9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.915s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1365.779424] env[62385]: DEBUG nova.compute.manager [req-583fa19c-5f71-4722-8a7c-e9f0003679e6 req-26e731d1-6a1e-42e0-ab5f-ebb18274711d service nova] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Received event network-changed-debdc145-742b-4723-b7a0-f34cfa0f749d {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1365.779624] env[62385]: DEBUG nova.compute.manager [req-583fa19c-5f71-4722-8a7c-e9f0003679e6 req-26e731d1-6a1e-42e0-ab5f-ebb18274711d service nova] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Refreshing instance network info cache due to event network-changed-debdc145-742b-4723-b7a0-f34cfa0f749d. {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1365.779850] env[62385]: DEBUG oslo_concurrency.lockutils [req-583fa19c-5f71-4722-8a7c-e9f0003679e6 req-26e731d1-6a1e-42e0-ab5f-ebb18274711d service nova] Acquiring lock "refresh_cache-0793179a-ce76-4404-a036-a72b6dac82e9" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1365.779998] env[62385]: DEBUG oslo_concurrency.lockutils [req-583fa19c-5f71-4722-8a7c-e9f0003679e6 req-26e731d1-6a1e-42e0-ab5f-ebb18274711d service nova] Acquired lock "refresh_cache-0793179a-ce76-4404-a036-a72b6dac82e9" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1365.780182] env[62385]: DEBUG nova.network.neutron [req-583fa19c-5f71-4722-8a7c-e9f0003679e6 req-26e731d1-6a1e-42e0-ab5f-ebb18274711d service nova] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Refreshing network info cache for port debdc145-742b-4723-b7a0-f34cfa0f749d {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1366.025453] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1366.029008] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1366.029196] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1366.029343] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62385) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1366.029502] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1366.029627] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Cleaning up deleted instances with incomplete migration {{(pid=62385) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1366.491434] env[62385]: DEBUG nova.network.neutron [req-583fa19c-5f71-4722-8a7c-e9f0003679e6 req-26e731d1-6a1e-42e0-ab5f-ebb18274711d service nova] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Updated VIF entry in instance network info cache for port debdc145-742b-4723-b7a0-f34cfa0f749d. {{(pid=62385) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1366.491803] env[62385]: DEBUG nova.network.neutron [req-583fa19c-5f71-4722-8a7c-e9f0003679e6 req-26e731d1-6a1e-42e0-ab5f-ebb18274711d service nova] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Updating instance_info_cache with network_info: [{"id": "debdc145-742b-4723-b7a0-f34cfa0f749d", "address": "fa:16:3e:58:ee:91", "network": {"id": "5aef0aec-abdf-4636-9d6b-ba476c7f0410", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1388512761-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.216", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4831e1f029ed4ffc8ce5bfe7d3f38dd3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "62f28d75-4e6a-4ae5-b8b3-d0652ea26d08", "external-id": "nsx-vlan-transportzone-801", "segmentation_id": 801, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdebdc145-74", "ovs_interfaceid": "debdc145-742b-4723-b7a0-f34cfa0f749d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1366.994324] env[62385]: DEBUG oslo_concurrency.lockutils [req-583fa19c-5f71-4722-8a7c-e9f0003679e6 req-26e731d1-6a1e-42e0-ab5f-ebb18274711d service nova] Releasing lock "refresh_cache-0793179a-ce76-4404-a036-a72b6dac82e9" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1368.533142] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1368.533526] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Starting heal instance info cache {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1368.533526] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Rebuilding the list of instances to heal {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1369.063184] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquiring lock "refresh_cache-0793179a-ce76-4404-a036-a72b6dac82e9" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1369.063339] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquired lock "refresh_cache-0793179a-ce76-4404-a036-a72b6dac82e9" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1369.063483] env[62385]: DEBUG nova.network.neutron [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Forcefully refreshing network info cache for instance {{(pid=62385) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1369.063634] env[62385]: DEBUG nova.objects.instance [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lazy-loading 'info_cache' on Instance uuid 0793179a-ce76-4404-a036-a72b6dac82e9 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1370.769805] env[62385]: DEBUG nova.network.neutron [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Updating instance_info_cache with network_info: [{"id": "debdc145-742b-4723-b7a0-f34cfa0f749d", "address": "fa:16:3e:58:ee:91", "network": {"id": "5aef0aec-abdf-4636-9d6b-ba476c7f0410", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1388512761-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.216", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4831e1f029ed4ffc8ce5bfe7d3f38dd3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "62f28d75-4e6a-4ae5-b8b3-d0652ea26d08", "external-id": "nsx-vlan-transportzone-801", "segmentation_id": 801, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdebdc145-74", "ovs_interfaceid": "debdc145-742b-4723-b7a0-f34cfa0f749d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1371.272470] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Releasing lock "refresh_cache-0793179a-ce76-4404-a036-a72b6dac82e9" {{(pid=62385) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1371.272710] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Updated the network info_cache for instance {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1371.273128] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1371.274017] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1371.274017] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager.update_available_resource {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1371.776859] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1371.777250] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1371.777250] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1371.777475] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62385) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1371.778671] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f63481a-ed12-4fdb-a5d2-effae61b5fe9 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1371.786690] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc994339-abad-4d9e-877a-27eb43b2a1cb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1371.800710] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-893795c2-40c8-441e-9df6-672d3253dd30 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1371.807655] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd52ca9e-b15e-4fff-92d6-13c960a4e73b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1371.838194] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181199MB free_disk=155GB free_vcpus=48 pci_devices=None {{(pid=62385) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1371.838369] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1371.838546] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1372.926317] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Instance 0793179a-ce76-4404-a036-a72b6dac82e9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62385) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1372.926583] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=62385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1372.926675] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=62385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1372.950947] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3624a722-d407-4adc-a0ae-da0d27d98851 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1372.958623] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-567e4f45-3a87-4d63-9bb3-195afb363b70 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1372.986771] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-612cc20b-104e-404c-8d42-efe45d1a0ecc {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1372.993222] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89a87904-94c3-41ea-ba50-b741de4a7fd2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1373.006497] env[62385]: DEBUG nova.compute.provider_tree [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1373.509758] env[62385]: DEBUG nova.scheduler.client.report [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1374.014940] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62385) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1374.015320] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.177s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1374.015399] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1374.015535] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Cleaning up deleted instances {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1374.523218] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] There are 16 instances to clean {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1374.523402] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: a600e434-fb16-4881-acb9-e5a55c03f152] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1375.027174] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: aa7fdda1-3441-4bc2-b2a3-c27ce91baab9] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1375.530562] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: a6a6c7de-656d-4358-977b-f62b471664da] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1376.034604] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 28afc06e-f6e4-4044-8bfe-c93b1d9c1388] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1376.538160] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 4fe375be-ba60-4601-bd89-610e12742df3] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1377.041194] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: c4f0482d-5b46-4c28-9c40-a59b6aee7dc3] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1377.544260] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 2cb0fee4-c922-402a-9251-10dc0da8c87d] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1378.048244] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: f93d9f5d-f7b6-4988-af7c-877a9e4c2104] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1378.551853] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 5c9284ba-9f83-454f-b91b-9b0ea9ee3195] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1379.055286] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 1289e63a-5a9d-4615-b6cf-bbf5f4ae0317] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1379.559256] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: c49d4366-2ea0-44f2-8463-daf62bd1c40c] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1380.062648] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 5a5bcb7c-86cd-4950-a3dd-15515ae0aca6] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1380.565865] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 68748d25-017f-4b58-9984-70264abbc5ba] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1381.068947] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 57637d6e-39e2-405c-bbcf-64bf0ffbebd1] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1381.572839] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 8e1d416c-2624-468c-94d7-6265e9f4178f] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1382.076615] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] [instance: 9d6f098a-0b05-43ef-96b0-9eb99ad3538c] Instance has had 0 of 5 cleanup attempts {{(pid=62385) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1382.579829] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1404.346174] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8f703c77-e993-4c16-9af0-1fbc52042e97 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "0793179a-ce76-4404-a036-a72b6dac82e9" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1404.346462] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8f703c77-e993-4c16-9af0-1fbc52042e97 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "0793179a-ce76-4404-a036-a72b6dac82e9" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1404.850081] env[62385]: DEBUG nova.compute.utils [None req-8f703c77-e993-4c16-9af0-1fbc52042e97 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Using /dev/sd instead of None {{(pid=62385) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1405.353653] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8f703c77-e993-4c16-9af0-1fbc52042e97 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "0793179a-ce76-4404-a036-a72b6dac82e9" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1406.411250] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8f703c77-e993-4c16-9af0-1fbc52042e97 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "0793179a-ce76-4404-a036-a72b6dac82e9" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1406.411683] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8f703c77-e993-4c16-9af0-1fbc52042e97 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "0793179a-ce76-4404-a036-a72b6dac82e9" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1406.411789] env[62385]: INFO nova.compute.manager [None req-8f703c77-e993-4c16-9af0-1fbc52042e97 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Attaching volume f6cc5d6b-3db2-4963-a78c-0ece7a5ae866 to /dev/sdb [ 1406.441010] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64da83a7-37df-45ca-9d06-4cf2237c92dd {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1406.447989] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-354fe025-8da1-4e5f-ae55-1196fe81448f {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1406.460453] env[62385]: DEBUG nova.virt.block_device [None req-8f703c77-e993-4c16-9af0-1fbc52042e97 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Updating existing volume attachment record: 3bfb9e2d-0769-4849-ba8f-5adca8f15f79 {{(pid=62385) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1411.003465] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f703c77-e993-4c16-9af0-1fbc52042e97 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Volume attach. Driver type: vmdk {{(pid=62385) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1411.003742] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f703c77-e993-4c16-9af0-1fbc52042e97 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261332', 'volume_id': 'f6cc5d6b-3db2-4963-a78c-0ece7a5ae866', 'name': 'volume-f6cc5d6b-3db2-4963-a78c-0ece7a5ae866', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0793179a-ce76-4404-a036-a72b6dac82e9', 'attached_at': '', 'detached_at': '', 'volume_id': 'f6cc5d6b-3db2-4963-a78c-0ece7a5ae866', 'serial': 'f6cc5d6b-3db2-4963-a78c-0ece7a5ae866'} {{(pid=62385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1411.004664] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b0a4b37-85cc-4d67-909e-99bbc5753b8c {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1411.020842] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39c28879-5d05-441d-956b-d1768914f00a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1411.044162] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f703c77-e993-4c16-9af0-1fbc52042e97 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Reconfiguring VM instance instance-00000076 to attach disk [datastore2] volume-f6cc5d6b-3db2-4963-a78c-0ece7a5ae866/volume-f6cc5d6b-3db2-4963-a78c-0ece7a5ae866.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1411.044390] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d5953ff0-f6ce-44bb-8a03-9c7d2a07aea2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1411.061659] env[62385]: DEBUG oslo_vmware.api [None req-8f703c77-e993-4c16-9af0-1fbc52042e97 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1411.061659] env[62385]: value = "task-1206631" [ 1411.061659] env[62385]: _type = "Task" [ 1411.061659] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1411.069323] env[62385]: DEBUG oslo_vmware.api [None req-8f703c77-e993-4c16-9af0-1fbc52042e97 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206631, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1411.571385] env[62385]: DEBUG oslo_vmware.api [None req-8f703c77-e993-4c16-9af0-1fbc52042e97 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206631, 'name': ReconfigVM_Task, 'duration_secs': 0.356044} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1411.571732] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f703c77-e993-4c16-9af0-1fbc52042e97 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Reconfigured VM instance instance-00000076 to attach disk [datastore2] volume-f6cc5d6b-3db2-4963-a78c-0ece7a5ae866/volume-f6cc5d6b-3db2-4963-a78c-0ece7a5ae866.vmdk or device None with type thin {{(pid=62385) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1411.576400] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5a0b7092-93f3-43d8-a3f2-58108858f55e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1411.591762] env[62385]: DEBUG oslo_vmware.api [None req-8f703c77-e993-4c16-9af0-1fbc52042e97 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1411.591762] env[62385]: value = "task-1206632" [ 1411.591762] env[62385]: _type = "Task" [ 1411.591762] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1411.599245] env[62385]: DEBUG oslo_vmware.api [None req-8f703c77-e993-4c16-9af0-1fbc52042e97 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206632, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1412.101882] env[62385]: DEBUG oslo_vmware.api [None req-8f703c77-e993-4c16-9af0-1fbc52042e97 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206632, 'name': ReconfigVM_Task, 'duration_secs': 0.15095} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1412.102205] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f703c77-e993-4c16-9af0-1fbc52042e97 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261332', 'volume_id': 'f6cc5d6b-3db2-4963-a78c-0ece7a5ae866', 'name': 'volume-f6cc5d6b-3db2-4963-a78c-0ece7a5ae866', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0793179a-ce76-4404-a036-a72b6dac82e9', 'attached_at': '', 'detached_at': '', 'volume_id': 'f6cc5d6b-3db2-4963-a78c-0ece7a5ae866', 'serial': 'f6cc5d6b-3db2-4963-a78c-0ece7a5ae866'} {{(pid=62385) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1413.160554] env[62385]: DEBUG nova.objects.instance [None req-8f703c77-e993-4c16-9af0-1fbc52042e97 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lazy-loading 'flavor' on Instance uuid 0793179a-ce76-4404-a036-a72b6dac82e9 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1413.666995] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8f703c77-e993-4c16-9af0-1fbc52042e97 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "0793179a-ce76-4404-a036-a72b6dac82e9" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.255s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1413.837360] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8a0b909e-97e0-4529-b077-044a62e93072 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "0793179a-ce76-4404-a036-a72b6dac82e9" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1413.837629] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8a0b909e-97e0-4529-b077-044a62e93072 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "0793179a-ce76-4404-a036-a72b6dac82e9" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1414.340349] env[62385]: INFO nova.compute.manager [None req-8a0b909e-97e0-4529-b077-044a62e93072 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Detaching volume f6cc5d6b-3db2-4963-a78c-0ece7a5ae866 [ 1414.370327] env[62385]: INFO nova.virt.block_device [None req-8a0b909e-97e0-4529-b077-044a62e93072 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Attempting to driver detach volume f6cc5d6b-3db2-4963-a78c-0ece7a5ae866 from mountpoint /dev/sdb [ 1414.370560] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a0b909e-97e0-4529-b077-044a62e93072 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Volume detach. Driver type: vmdk {{(pid=62385) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1414.370772] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a0b909e-97e0-4529-b077-044a62e93072 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261332', 'volume_id': 'f6cc5d6b-3db2-4963-a78c-0ece7a5ae866', 'name': 'volume-f6cc5d6b-3db2-4963-a78c-0ece7a5ae866', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0793179a-ce76-4404-a036-a72b6dac82e9', 'attached_at': '', 'detached_at': '', 'volume_id': 'f6cc5d6b-3db2-4963-a78c-0ece7a5ae866', 'serial': 'f6cc5d6b-3db2-4963-a78c-0ece7a5ae866'} {{(pid=62385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1414.371625] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15097b7f-c002-43db-8fe9-00197202493a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1414.392816] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b52cc21-5a10-49ce-a737-35c8a2f30e5b {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1414.399252] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a428d42-f166-4d28-8a24-ec8aa8f3bd87 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1414.418371] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b51b9146-f73a-4e00-803d-e0125cdee810 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1414.432499] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a0b909e-97e0-4529-b077-044a62e93072 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] The volume has not been displaced from its original location: [datastore2] volume-f6cc5d6b-3db2-4963-a78c-0ece7a5ae866/volume-f6cc5d6b-3db2-4963-a78c-0ece7a5ae866.vmdk. No consolidation needed. {{(pid=62385) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1414.437719] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a0b909e-97e0-4529-b077-044a62e93072 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Reconfiguring VM instance instance-00000076 to detach disk 2001 {{(pid=62385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1414.437974] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cf6300d6-94ca-4761-b7c9-73c7bdbe03e3 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1414.455299] env[62385]: DEBUG oslo_vmware.api [None req-8a0b909e-97e0-4529-b077-044a62e93072 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1414.455299] env[62385]: value = "task-1206633" [ 1414.455299] env[62385]: _type = "Task" [ 1414.455299] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1414.462374] env[62385]: DEBUG oslo_vmware.api [None req-8a0b909e-97e0-4529-b077-044a62e93072 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206633, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1414.964493] env[62385]: DEBUG oslo_vmware.api [None req-8a0b909e-97e0-4529-b077-044a62e93072 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206633, 'name': ReconfigVM_Task, 'duration_secs': 0.215173} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1414.964770] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a0b909e-97e0-4529-b077-044a62e93072 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Reconfigured VM instance instance-00000076 to detach disk 2001 {{(pid=62385) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1414.969295] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e9a56625-dd59-41e3-b674-83732e7eb97a {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1414.983464] env[62385]: DEBUG oslo_vmware.api [None req-8a0b909e-97e0-4529-b077-044a62e93072 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1414.983464] env[62385]: value = "task-1206634" [ 1414.983464] env[62385]: _type = "Task" [ 1414.983464] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1414.990636] env[62385]: DEBUG oslo_vmware.api [None req-8a0b909e-97e0-4529-b077-044a62e93072 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206634, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1415.494236] env[62385]: DEBUG oslo_vmware.api [None req-8a0b909e-97e0-4529-b077-044a62e93072 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206634, 'name': ReconfigVM_Task, 'duration_secs': 0.130449} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1415.494685] env[62385]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a0b909e-97e0-4529-b077-044a62e93072 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-261332', 'volume_id': 'f6cc5d6b-3db2-4963-a78c-0ece7a5ae866', 'name': 'volume-f6cc5d6b-3db2-4963-a78c-0ece7a5ae866', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0793179a-ce76-4404-a036-a72b6dac82e9', 'attached_at': '', 'detached_at': '', 'volume_id': 'f6cc5d6b-3db2-4963-a78c-0ece7a5ae866', 'serial': 'f6cc5d6b-3db2-4963-a78c-0ece7a5ae866'} {{(pid=62385) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1416.035299] env[62385]: DEBUG nova.objects.instance [None req-8a0b909e-97e0-4529-b077-044a62e93072 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lazy-loading 'flavor' on Instance uuid 0793179a-ce76-4404-a036-a72b6dac82e9 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1417.041949] env[62385]: DEBUG oslo_concurrency.lockutils [None req-8a0b909e-97e0-4529-b077-044a62e93072 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "0793179a-ce76-4404-a036-a72b6dac82e9" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.204s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1418.071908] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ff596b34-cae0-4a76-8fd0-6d040ff12eb3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "0793179a-ce76-4404-a036-a72b6dac82e9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1418.072298] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ff596b34-cae0-4a76-8fd0-6d040ff12eb3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "0793179a-ce76-4404-a036-a72b6dac82e9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1418.072441] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ff596b34-cae0-4a76-8fd0-6d040ff12eb3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "0793179a-ce76-4404-a036-a72b6dac82e9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1418.072635] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ff596b34-cae0-4a76-8fd0-6d040ff12eb3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "0793179a-ce76-4404-a036-a72b6dac82e9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1418.072812] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ff596b34-cae0-4a76-8fd0-6d040ff12eb3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "0793179a-ce76-4404-a036-a72b6dac82e9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1418.074944] env[62385]: INFO nova.compute.manager [None req-ff596b34-cae0-4a76-8fd0-6d040ff12eb3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Terminating instance [ 1418.076793] env[62385]: DEBUG nova.compute.manager [None req-ff596b34-cae0-4a76-8fd0-6d040ff12eb3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Start destroying the instance on the hypervisor. {{(pid=62385) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1418.076986] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ff596b34-cae0-4a76-8fd0-6d040ff12eb3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Destroying instance {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1418.077803] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75c9c718-1c49-47ce-b833-f42563b7bb85 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1418.085473] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff596b34-cae0-4a76-8fd0-6d040ff12eb3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Powering off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1418.085700] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-101b1f4d-5c77-4269-9874-0ca700a890f7 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1418.092337] env[62385]: DEBUG oslo_vmware.api [None req-ff596b34-cae0-4a76-8fd0-6d040ff12eb3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1418.092337] env[62385]: value = "task-1206635" [ 1418.092337] env[62385]: _type = "Task" [ 1418.092337] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1418.100453] env[62385]: DEBUG oslo_vmware.api [None req-ff596b34-cae0-4a76-8fd0-6d040ff12eb3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206635, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1418.602378] env[62385]: DEBUG oslo_vmware.api [None req-ff596b34-cae0-4a76-8fd0-6d040ff12eb3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206635, 'name': PowerOffVM_Task, 'duration_secs': 0.175453} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1418.602629] env[62385]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff596b34-cae0-4a76-8fd0-6d040ff12eb3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Powered off the VM {{(pid=62385) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1418.602801] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ff596b34-cae0-4a76-8fd0-6d040ff12eb3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Unregistering the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1418.603068] env[62385]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e09164c6-82ff-487a-9866-abf0eb05d231 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1418.665860] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ff596b34-cae0-4a76-8fd0-6d040ff12eb3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Unregistered the VM {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1418.666064] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ff596b34-cae0-4a76-8fd0-6d040ff12eb3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Deleting contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1418.666217] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff596b34-cae0-4a76-8fd0-6d040ff12eb3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Deleting the datastore file [datastore1] 0793179a-ce76-4404-a036-a72b6dac82e9 {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1418.666475] env[62385]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c5462ced-44f9-43c2-8f91-ddcda3ff35be {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1418.673353] env[62385]: DEBUG oslo_vmware.api [None req-ff596b34-cae0-4a76-8fd0-6d040ff12eb3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for the task: (returnval){ [ 1418.673353] env[62385]: value = "task-1206637" [ 1418.673353] env[62385]: _type = "Task" [ 1418.673353] env[62385]: } to complete. {{(pid=62385) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1418.680712] env[62385]: DEBUG oslo_vmware.api [None req-ff596b34-cae0-4a76-8fd0-6d040ff12eb3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206637, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1419.183435] env[62385]: DEBUG oslo_vmware.api [None req-ff596b34-cae0-4a76-8fd0-6d040ff12eb3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Task: {'id': task-1206637, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.128128} completed successfully. {{(pid=62385) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1419.183824] env[62385]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff596b34-cae0-4a76-8fd0-6d040ff12eb3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Deleted the datastore file {{(pid=62385) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1419.183967] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ff596b34-cae0-4a76-8fd0-6d040ff12eb3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Deleted contents of the VM from datastore datastore1 {{(pid=62385) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1419.184164] env[62385]: DEBUG nova.virt.vmwareapi.vmops [None req-ff596b34-cae0-4a76-8fd0-6d040ff12eb3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Instance destroyed {{(pid=62385) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1419.184348] env[62385]: INFO nova.compute.manager [None req-ff596b34-cae0-4a76-8fd0-6d040ff12eb3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1419.184590] env[62385]: DEBUG oslo.service.loopingcall [None req-ff596b34-cae0-4a76-8fd0-6d040ff12eb3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62385) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1419.184779] env[62385]: DEBUG nova.compute.manager [-] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Deallocating network for instance {{(pid=62385) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1419.184874] env[62385]: DEBUG nova.network.neutron [-] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] deallocate_for_instance() {{(pid=62385) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1419.591109] env[62385]: DEBUG nova.compute.manager [req-5389661a-06ae-449d-9473-390c2050fe85 req-6c97bc98-94f2-45da-a62f-f6d3104d94d2 service nova] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Received event network-vif-deleted-debdc145-742b-4723-b7a0-f34cfa0f749d {{(pid=62385) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1419.591250] env[62385]: INFO nova.compute.manager [req-5389661a-06ae-449d-9473-390c2050fe85 req-6c97bc98-94f2-45da-a62f-f6d3104d94d2 service nova] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Neutron deleted interface debdc145-742b-4723-b7a0-f34cfa0f749d; detaching it from the instance and deleting it from the info cache [ 1419.591408] env[62385]: DEBUG nova.network.neutron [req-5389661a-06ae-449d-9473-390c2050fe85 req-6c97bc98-94f2-45da-a62f-f6d3104d94d2 service nova] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1420.071174] env[62385]: DEBUG nova.network.neutron [-] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Updating instance_info_cache with network_info: [] {{(pid=62385) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1420.093744] env[62385]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-632e3686-698c-407f-9aec-a1719df735fb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1420.104581] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39539fdd-bfcc-48e3-aaca-d45e62c62032 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1420.126490] env[62385]: DEBUG nova.compute.manager [req-5389661a-06ae-449d-9473-390c2050fe85 req-6c97bc98-94f2-45da-a62f-f6d3104d94d2 service nova] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Detach interface failed, port_id=debdc145-742b-4723-b7a0-f34cfa0f749d, reason: Instance 0793179a-ce76-4404-a036-a72b6dac82e9 could not be found. {{(pid=62385) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1420.574101] env[62385]: INFO nova.compute.manager [-] [instance: 0793179a-ce76-4404-a036-a72b6dac82e9] Took 1.39 seconds to deallocate network for instance. [ 1421.081202] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ff596b34-cae0-4a76-8fd0-6d040ff12eb3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1421.081487] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ff596b34-cae0-4a76-8fd0-6d040ff12eb3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1421.081720] env[62385]: DEBUG nova.objects.instance [None req-ff596b34-cae0-4a76-8fd0-6d040ff12eb3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lazy-loading 'resources' on Instance uuid 0793179a-ce76-4404-a036-a72b6dac82e9 {{(pid=62385) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1421.614535] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a405b71-e76b-4ddd-96a5-04adb1423033 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1421.622176] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5e25aa2-0a33-4edf-9079-1c7b29757406 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1421.650551] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3094f862-6e29-44c3-8ecb-a767a8c612b2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1421.658209] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecb49d00-fdc2-4832-aad5-2bada5530951 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1421.671150] env[62385]: DEBUG nova.compute.provider_tree [None req-ff596b34-cae0-4a76-8fd0-6d040ff12eb3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1422.174189] env[62385]: DEBUG nova.scheduler.client.report [None req-ff596b34-cae0-4a76-8fd0-6d040ff12eb3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1422.680588] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ff596b34-cae0-4a76-8fd0-6d040ff12eb3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.599s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1422.699851] env[62385]: INFO nova.scheduler.client.report [None req-ff596b34-cae0-4a76-8fd0-6d040ff12eb3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Deleted allocations for instance 0793179a-ce76-4404-a036-a72b6dac82e9 [ 1423.207867] env[62385]: DEBUG oslo_concurrency.lockutils [None req-ff596b34-cae0-4a76-8fd0-6d040ff12eb3 tempest-AttachVolumeNegativeTest-173497814 tempest-AttachVolumeNegativeTest-173497814-project-member] Lock "0793179a-ce76-4404-a036-a72b6dac82e9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.136s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1436.839574] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1436.840038] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1437.344076] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1437.344076] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Starting heal instance info cache {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1437.344076] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Rebuilding the list of instances to heal {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1437.846250] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Didn't find any instances for network info cache update. {{(pid=62385) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1437.846748] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1437.846748] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1437.846748] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1437.846918] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1437.847048] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1437.847194] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1437.847320] env[62385]: DEBUG nova.compute.manager [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62385) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1437.847459] env[62385]: DEBUG oslo_service.periodic_task [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Running periodic task ComputeManager.update_available_resource {{(pid=62385) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1438.350791] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1438.351060] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1438.351217] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1438.351376] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62385) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1438.352309] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1219022a-50c2-4033-9ded-c7e95f188a63 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1438.360824] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b409974f-f877-4907-b50a-034b3664a9c1 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1438.374475] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94e24aae-7527-423a-8638-de0a2c3ed37e {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1438.380633] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1352ba3-199c-4469-b405-cf4c0a95d904 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1438.408754] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181194MB free_disk=155GB free_vcpus=48 pci_devices=None {{(pid=62385) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1438.408936] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1438.409160] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1439.428151] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1439.428408] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62385) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1439.441447] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb2ebe1a-3f04-4676-bf7b-27867622e250 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1439.449018] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-073bc965-3b12-4924-af7d-c248250e4fdb {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1439.477612] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecd07c65-5161-4da8-862d-b1943924e2f2 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1439.484856] env[62385]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80be50c2-1b97-44dc-a9d5-36f33a03f879 {{(pid=62385) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1439.498109] env[62385]: DEBUG nova.compute.provider_tree [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Inventory has not changed in ProviderTree for provider: 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a {{(pid=62385) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1440.001360] env[62385]: DEBUG nova.scheduler.client.report [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Inventory has not changed for provider 8c6086eb-4f96-48ed-af58-d7d7f36e7c5a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 155, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62385) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1440.505380] env[62385]: DEBUG nova.compute.resource_tracker [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62385) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1440.505748] env[62385]: DEBUG oslo_concurrency.lockutils [None req-879d669b-441e-4872-b614-9e2bbf62ea0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.096s {{(pid=62385) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}}